diff --git a/src/test/regress/expected/multi_array_agg.out b/src/test/regress/expected/multi_array_agg.out index b3fe1b307..b473b7967 100644 --- a/src/test/regress/expected/multi_array_agg.out +++ b/src/test/regress/expected/multi_array_agg.out @@ -1,6 +1,8 @@ -- -- MULTI_ARRAY_AGG -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 520000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 520000; -- Check multi_cat_agg() aggregate which is used to implement array_agg() SELECT array_cat_agg(i) FROM (VALUES (ARRAY[1,2]), (NULL), (ARRAY[3,4])) AS t(i); array_cat_agg @@ -96,10 +98,10 @@ SELECT l_quantity, count(*), avg(l_extendedprice), array_agg(l_orderkey) FROM li GROUP BY l_quantity ORDER BY l_quantity; l_quantity | count | avg | array_agg ------------+-------+-----------------------+-------------------------------------------------------------------------------------------------- - 1.00 | 17 | 1477.1258823529411765 | {8997,9026,9158,9184,9220,9222,9348,9383,9476,5543,5633,5634,5698,5766,5856,5857,5986} - 2.00 | 19 | 3078.4242105263157895 | {9030,9058,9123,9124,9188,9344,9441,9476,5506,5540,5573,5669,5703,5730,5798,5831,5893,5920,5923} - 3.00 | 14 | 4714.0392857142857143 | {9124,9157,9184,9223,9254,9349,9414,9475,9477,5509,5543,5605,5606,5827} - 4.00 | 19 | 5929.7136842105263158 | {9091,9120,9281,9347,9382,9440,9473,5504,5507,5508,5511,5538,5764,5766,5826,5829,5862,5959,5985} + 1.00 | 17 | 1477.1258823529411765 | {5543,5633,5634,5698,5766,5856,5857,5986,8997,9026,9158,9184,9220,9222,9348,9383,9476} + 2.00 | 19 | 3078.4242105263157895 | {5506,5540,5573,5669,5703,5730,5798,5831,5893,5920,5923,9030,9058,9123,9124,9188,9344,9441,9476} + 3.00 | 14 | 4714.0392857142857143 | {5509,5543,5605,5606,5827,9124,9157,9184,9223,9254,9349,9414,9475,9477} + 4.00 | 19 | 5929.7136842105263158 | {5504,5507,5508,5511,5538,5764,5766,5826,5829,5862,5959,5985,9091,9120,9281,9347,9382,9440,9473} (4 rows) SELECT l_quantity, array_agg(extract (month FROM o_orderdate)) AS my_month @@ -107,10 +109,10 @@ SELECT l_quantity, array_agg(extract (month FROM o_orderdate)) AS my_month AND l_orderkey > 5500 AND l_orderkey < 9500 GROUP BY l_quantity ORDER BY l_quantity; l_quantity | my_month ------------+------------------------------------------------ - 1.00 | {7,7,4,7,4,2,6,3,5,9,5,7,5,9,11,11,4} - 2.00 | {7,6,6,10,1,12,6,5,11,10,8,5,5,12,3,11,7,11,5} - 3.00 | {10,6,7,8,5,8,9,11,3,4,9,8,11,7} - 4.00 | {11,6,2,8,2,6,10,1,5,6,11,12,10,9,6,1,2,5,1} + 1.00 | {9,5,7,5,9,11,11,4,7,7,4,7,4,2,6,3,5} + 2.00 | {11,10,8,5,5,12,3,11,7,11,5,7,6,6,10,1,12,6,5} + 3.00 | {4,9,8,11,7,10,6,7,8,5,8,9,11,3} + 4.00 | {1,5,6,11,12,10,9,6,1,2,5,1,11,6,2,8,2,6,10} (4 rows) SELECT l_quantity, array_agg(l_orderkey * 2 + 1) FROM lineitem WHERE l_quantity < 5 @@ -118,10 +120,10 @@ SELECT l_quantity, array_agg(l_orderkey * 2 + 1) FROM lineitem WHERE l_quantity AND l_orderkey > 5500 AND l_orderkey < 9500 GROUP BY l_quantity ORDER BY l_quantity; l_quantity | array_agg ------------+--------------------------------------------- - 1.00 | {18317,18445,11269,11397,11713,11715,11973} - 2.00 | {18061,18247,18953,11847} + 1.00 | {11269,11397,11713,11715,11973,18317,18445} + 2.00 | {11847,18061,18247,18953} 3.00 | {18249,18315,18699,18951,18955} - 4.00 | {18241,18765,11653,11659} + 4.00 | {11653,11659,18241,18765} (4 rows) -- Check that we can execute array_agg() with an expression containing NULL values diff --git a/src/test/regress/expected/multi_average_expression.out b/src/test/regress/expected/multi_average_expression.out index 74511d553..a0df52dfc 100644 --- a/src/test/regress/expected/multi_average_expression.out +++ b/src/test/regress/expected/multi_average_expression.out @@ -4,6 +4,8 @@ -- This test checks that the group-by columns don't need to be above an average -- expression, and can be anywhere in the projection order. This is in response -- to a bug we had due to the average expression introducing new columns. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 450000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 450000; SELECT sum(l_quantity) as sum_qty, sum(l_extendedprice) as sum_base_price, diff --git a/src/test/regress/expected/multi_basic_queries.out b/src/test/regress/expected/multi_basic_queries.out index f226ac6e3..5a3eaaf04 100644 --- a/src/test/regress/expected/multi_basic_queries.out +++ b/src/test/regress/expected/multi_basic_queries.out @@ -1,6 +1,8 @@ -- -- MULTI_BASIC_QUERIES -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 440000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 440000; -- Execute simple sum, average, and count queries on data recently uploaded to -- our partitioned table. SELECT count(*) FROM lineitem; diff --git a/src/test/regress/expected/multi_binary_master_copy_format.out b/src/test/regress/expected/multi_binary_master_copy_format.out index 505d18a34..5e213e175 100644 --- a/src/test/regress/expected/multi_binary_master_copy_format.out +++ b/src/test/regress/expected/multi_binary_master_copy_format.out @@ -1,6 +1,8 @@ -- -- MULTI_BINARY_MASTER_COPY -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 430000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 430000; -- Try binary master copy for different executors SET citus.binary_master_copy_format TO 'on'; SET citus.task_executor_type TO 'task-tracker'; diff --git a/src/test/regress/expected/multi_complex_expressions.out b/src/test/regress/expected/multi_complex_expressions.out index f1f902586..3217e1681 100644 --- a/src/test/regress/expected/multi_complex_expressions.out +++ b/src/test/regress/expected/multi_complex_expressions.out @@ -1,6 +1,8 @@ -- -- MULTI_COMPLEX_EXPRESSIONS -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 420000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 420000; -- Check that we can correctly handle complex expressions and aggregates. SELECT sum(l_quantity) / avg(l_quantity) FROM lineitem; ?column? diff --git a/src/test/regress/expected/multi_connection_cache.out b/src/test/regress/expected/multi_connection_cache.out index ac808a0af..84e927bbe 100644 --- a/src/test/regress/expected/multi_connection_cache.out +++ b/src/test/regress/expected/multi_connection_cache.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 410000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 410000; -- =================================================================== -- create test functions -- =================================================================== diff --git a/src/test/regress/expected/multi_count_type_conversion.out b/src/test/regress/expected/multi_count_type_conversion.out index a258ebef2..ef5cefef6 100644 --- a/src/test/regress/expected/multi_count_type_conversion.out +++ b/src/test/regress/expected/multi_count_type_conversion.out @@ -1,6 +1,8 @@ -- -- MULTI_COUNT_TYPE_CONVERSION -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 400000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 400000; -- Verify that we can sort count(*) results correctly. We perform this check as -- our count() operations execute in two steps: worker nodes report their -- count() results, and the master node sums these counts up. During this sum(), diff --git a/src/test/regress/expected/multi_create_fdw.out b/src/test/regress/expected/multi_create_fdw.out index 006ad6d1a..c19446ab3 100644 --- a/src/test/regress/expected/multi_create_fdw.out +++ b/src/test/regress/expected/multi_create_fdw.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 390000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 390000; -- =================================================================== -- get ready for the foreign data wrapper tests -- =================================================================== diff --git a/src/test/regress/expected/multi_create_insert_proxy.out b/src/test/regress/expected/multi_create_insert_proxy.out index d90262c4a..93faa5e71 100644 --- a/src/test/regress/expected/multi_create_insert_proxy.out +++ b/src/test/regress/expected/multi_create_insert_proxy.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 380000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 380000; -- =================================================================== -- test INSERT proxy creation functionality -- =================================================================== diff --git a/src/test/regress/expected/multi_create_shards.out b/src/test/regress/expected/multi_create_shards.out index 39cb45eef..cf3dfc3cd 100644 --- a/src/test/regress/expected/multi_create_shards.out +++ b/src/test/regress/expected/multi_create_shards.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 370000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 370000; -- =================================================================== -- create test functions and types needed for tests -- =================================================================== diff --git a/src/test/regress/expected/multi_create_table.out b/src/test/regress/expected/multi_create_table.out index dd3f1c167..b768aebce 100644 --- a/src/test/regress/expected/multi_create_table.out +++ b/src/test/regress/expected/multi_create_table.out @@ -1,6 +1,8 @@ -- -- MULTI_CREATE_TABLE -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 360000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 360000; -- Create new table definitions for use in testing in distributed planning and -- execution functionality. Also create indexes to boost performance. CREATE TABLE lineitem ( diff --git a/src/test/regress/expected/multi_data_types.out b/src/test/regress/expected/multi_data_types.out index 18b7a54dc..2c3de9e53 100644 --- a/src/test/regress/expected/multi_data_types.out +++ b/src/test/regress/expected/multi_data_types.out @@ -2,6 +2,8 @@ -- test composite type, varchar and enum types -- create, distribute, INSERT, SELECT and UPDATE -- =================================================================== +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 530000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 530000; -- create a custom type... CREATE TYPE test_composite_type AS ( i integer, diff --git a/src/test/regress/expected/multi_distribution_metadata.out b/src/test/regress/expected/multi_distribution_metadata.out index 88d5a3ba1..5cc8c5473 100644 --- a/src/test/regress/expected/multi_distribution_metadata.out +++ b/src/test/regress/expected/multi_distribution_metadata.out @@ -1,6 +1,8 @@ -- =================================================================== -- create test functions -- =================================================================== +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 540000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 540000; CREATE FUNCTION load_shard_id_array(regclass) RETURNS bigint[] AS 'citus' @@ -75,16 +77,16 @@ SELECT master_create_worker_shards('events_hash', 4, 2); (1 row) -- set shardstate of one replication from each shard to 0 (invalid value) -UPDATE pg_dist_shard_placement SET shardstate = 0 WHERE nodeport = 57638 AND shardid BETWEEN 103025 AND 103028; +UPDATE pg_dist_shard_placement SET shardstate = 0 WHERE nodeport = 57638 AND shardid BETWEEN 540000 AND 540003; -- should see above shard identifiers SELECT load_shard_id_array('events_hash'); load_shard_id_array ------------------------------- - {103025,103026,103027,103028} + {540000,540001,540002,540003} (1 row) -- should see array with first shard range -SELECT load_shard_interval_array(103025, 0); +SELECT load_shard_interval_array(540000, 0); load_shard_interval_array --------------------------- {-2147483648,-1073741825} @@ -106,42 +108,41 @@ SELECT master_create_distributed_table('events_range', 'name', 'range'); SELECT master_create_empty_shard('events_range'); master_create_empty_shard --------------------------- - 103029 + 540004 (1 row) UPDATE pg_dist_shard SET shardminvalue = 'Aardvark', shardmaxvalue = 'Zebra' -WHERE shardid = 103029; -SELECT load_shard_interval_array(103029, ''::text); +WHERE shardid = 540004; +SELECT load_shard_interval_array(540004, ''::text); load_shard_interval_array --------------------------- {Aardvark,Zebra} (1 row) -- should see error for non-existent shard -SELECT load_shard_interval_array(103030, 0); -ERROR: could not find valid entry for shard 103030 +SELECT load_shard_interval_array(540005, 0); +ERROR: could not find valid entry for shard 540005 -- should see two placements -SELECT load_shard_placement_array(103026, false); +SELECT load_shard_placement_array(540001, false); load_shard_placement_array ----------------------------------- {localhost:57637,localhost:57638} (1 row) -- only one of which is finalized -SELECT load_shard_placement_array(103026, true); +SELECT load_shard_placement_array(540001, true); load_shard_placement_array ---------------------------- {localhost:57637} (1 row) -- should see error for non-existent shard -SELECT load_shard_placement_array(103031, false); -WARNING: could not find any shard placements for shardId 103031 - load_shard_placement_array ----------------------------- - {} +SELECT load_shard_placement_array(540001, false); + load_shard_placement_array +----------------------------------- + {localhost:57637,localhost:57638} (1 row) -- should see column id of 'name' @@ -192,7 +193,7 @@ SELECT column_name_to_column_id('events_hash', 'non_existent'); ERROR: column "non_existent" of relation "events_hash" does not exist -- drop shard rows (must drop placements first) DELETE FROM pg_dist_shard_placement - WHERE shardid BETWEEN 103025 AND 103029; + WHERE shardid BETWEEN 540000 AND 540004; DELETE FROM pg_dist_shard WHERE logicalrelid = 'events_hash'::regclass; DELETE FROM pg_dist_shard @@ -275,9 +276,9 @@ WHERE shardid = :new_shard_id AND nodename = 'localhost' and nodeport = 5432; -- deleting or updating a non-existent row should fail SELECT delete_shard_placement_row(:new_shard_id, 'wrong_localhost', 5432); -ERROR: could not find valid entry for shard placement 103030 on node "wrong_localhost:5432" +ERROR: could not find valid entry for shard placement 540005 on node "wrong_localhost:5432" SELECT update_shard_placement_row_state(:new_shard_id, 'localhost', 5432, 3); -ERROR: could not find valid entry for shard placement 103030 on node "localhost:5432" +ERROR: could not find valid entry for shard placement 540005 on node "localhost:5432" -- now we'll even test our lock methods... -- use transaction to bound how long we hold the lock BEGIN; diff --git a/src/test/regress/expected/multi_drop_extension.out b/src/test/regress/expected/multi_drop_extension.out index d5c1e2627..82bd8a47d 100644 --- a/src/test/regress/expected/multi_drop_extension.out +++ b/src/test/regress/expected/multi_drop_extension.out @@ -2,6 +2,8 @@ -- MULTI_DROP_EXTENSION -- -- Tests around dropping and recreating the extension +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 550000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 550000; CREATE TABLE testtableddl(somecol int, distributecol text NOT NULL); SELECT master_create_distributed_table('testtableddl', 'distributecol', 'append'); master_create_distributed_table diff --git a/src/test/regress/expected/multi_dropped_column_aliases.out b/src/test/regress/expected/multi_dropped_column_aliases.out index 4329adc63..e24ba9a5b 100644 --- a/src/test/regress/expected/multi_dropped_column_aliases.out +++ b/src/test/regress/expected/multi_dropped_column_aliases.out @@ -1,5 +1,7 @@ -- Tests that check that our query functionality behaves as expected when the -- table schema is modified via ALTER statements. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 620000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 620000; SELECT count(*) FROM customer; count ------- diff --git a/src/test/regress/expected/multi_explain.out b/src/test/regress/expected/multi_explain.out index 8ea35368a..870b77ae8 100644 --- a/src/test/regress/expected/multi_explain.out +++ b/src/test/regress/expected/multi_explain.out @@ -1,6 +1,8 @@ -- -- MULTI_EXPLAIN -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 570000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 570000; \a\t SET citus.task_executor_type TO 'real-time'; SET citus.explain_distributed_queries TO on; @@ -30,7 +32,7 @@ $BODY$ LANGUAGE plpgsql; EXPLAIN (COSTS FALSE, FORMAT TEXT) SELECT l_quantity, count(*) count_quantity FROM lineitem GROUP BY l_quantity ORDER BY count_quantity, l_quantity; -Distributed Query into pg_merge_job_0040 +Distributed Query into pg_merge_job_570000 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 @@ -38,13 +40,13 @@ Distributed Query into pg_merge_job_0040 Node: host=localhost port=57637 dbname=regression -> HashAggregate Group Key: l_quantity - -> Seq Scan on lineitem_102010 lineitem + -> Seq Scan on lineitem_290000 lineitem Master Query -> Sort - Sort Key: (sum(((sum(intermediate_column_40_1))::bigint)))::bigint, intermediate_column_40_0 + Sort Key: (sum(((sum(intermediate_column_570000_1))::bigint)))::bigint, intermediate_column_570000_0 -> HashAggregate - Group Key: intermediate_column_40_0 - -> Seq Scan on pg_merge_job_0040 + Group Key: intermediate_column_570000_0 + -> Seq Scan on pg_merge_job_570000 -- Test JSON format EXPLAIN (COSTS FALSE, FORMAT JSON) SELECT l_quantity, count(*) count_quantity FROM lineitem @@ -69,7 +71,7 @@ EXPLAIN (COSTS FALSE, FORMAT JSON) { "Node Type": "Seq Scan", "Parent Relationship": "Outer", - "Relation Name": "lineitem_102010", + "Relation Name": "lineitem_290000", "Alias": "lineitem" } ] @@ -85,19 +87,19 @@ EXPLAIN (COSTS FALSE, FORMAT JSON) { "Plan": { "Node Type": "Sort", - "Sort Key": ["(sum(((sum(intermediate_column_41_1))::bigint)))::bigint", "intermediate_column_41_0"], + "Sort Key": ["(sum(((sum(intermediate_column_570001_1))::bigint)))::bigint", "intermediate_column_570001_0"], "Plans": [ { "Node Type": "Aggregate", "Strategy": "Hashed", "Parent Relationship": "Outer", - "Group Key": ["intermediate_column_41_0"], + "Group Key": ["intermediate_column_570001_0"], "Plans": [ { "Node Type": "Seq Scan", "Parent Relationship": "Outer", - "Relation Name": "pg_merge_job_0041", - "Alias": "pg_merge_job_0041" + "Relation Name": "pg_merge_job_570001", + "Alias": "pg_merge_job_570001" } ] } @@ -138,7 +140,7 @@ EXPLAIN (COSTS FALSE, FORMAT XML) Seq Scan Outer - lineitem_102010 + lineitem_290000 lineitem @@ -154,8 +156,8 @@ EXPLAIN (COSTS FALSE, FORMAT XML) Sort - (sum(((sum(intermediate_column_43_1))::bigint)))::bigint - intermediate_column_43_0 + (sum(((sum(intermediate_column_570003_1))::bigint)))::bigint + intermediate_column_570003_0 @@ -163,14 +165,14 @@ EXPLAIN (COSTS FALSE, FORMAT XML) Hashed Outer - intermediate_column_43_0 + intermediate_column_570003_0 Seq Scan Outer - pg_merge_job_0043 - pg_merge_job_0043 + pg_merge_job_570003 + pg_merge_job_570003 @@ -204,31 +206,31 @@ EXPLAIN (COSTS FALSE, FORMAT YAML) Plans: - Node Type: "Seq Scan" Parent Relationship: "Outer" - Relation Name: "lineitem_102010" + Relation Name: "lineitem_290000" Alias: "lineitem" Master Query: - Plan: Node Type: "Sort" Sort Key: - - "(sum(((sum(intermediate_column_45_1))::bigint)))::bigint" - - "intermediate_column_45_0" + - "(sum(((sum(intermediate_column_570005_1))::bigint)))::bigint" + - "intermediate_column_570005_0" Plans: - Node Type: "Aggregate" Strategy: "Hashed" Parent Relationship: "Outer" Group Key: - - "intermediate_column_45_0" + - "intermediate_column_570005_0" Plans: - Node Type: "Seq Scan" Parent Relationship: "Outer" - Relation Name: "pg_merge_job_0045" - Alias: "pg_merge_job_0045" + Relation Name: "pg_merge_job_570005" + Alias: "pg_merge_job_570005" -- Test Text format EXPLAIN (COSTS FALSE, FORMAT TEXT) SELECT l_quantity, count(*) count_quantity FROM lineitem GROUP BY l_quantity ORDER BY count_quantity, l_quantity; -Distributed Query into pg_merge_job_0046 +Distributed Query into pg_merge_job_570006 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 @@ -236,17 +238,17 @@ Distributed Query into pg_merge_job_0046 Node: host=localhost port=57637 dbname=regression -> HashAggregate Group Key: l_quantity - -> Seq Scan on lineitem_102010 lineitem + -> Seq Scan on lineitem_290000 lineitem Master Query -> Sort - Sort Key: (sum(((sum(intermediate_column_46_1))::bigint)))::bigint, intermediate_column_46_0 + Sort Key: (sum(((sum(intermediate_column_570006_1))::bigint)))::bigint, intermediate_column_570006_0 -> HashAggregate - Group Key: intermediate_column_46_0 - -> Seq Scan on pg_merge_job_0046 + Group Key: intermediate_column_570006_0 + -> Seq Scan on pg_merge_job_570006 -- Test verbose EXPLAIN (COSTS FALSE, VERBOSE TRUE) SELECT sum(l_quantity) / avg(l_quantity) FROM lineitem; -Distributed Query into pg_merge_job_0047 +Distributed Query into pg_merge_job_570007 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 @@ -254,19 +256,19 @@ Distributed Query into pg_merge_job_0047 Node: host=localhost port=57637 dbname=regression -> Aggregate Output: sum(l_quantity), sum(l_quantity), count(l_quantity) - -> Seq Scan on public.lineitem_102010 lineitem + -> Seq Scan on public.lineitem_290000 lineitem Output: l_orderkey, l_partkey, l_suppkey, l_linenumber, l_quantity, l_extendedprice, l_discount, l_tax, l_returnflag, l_linestatus, l_shipdate, l_commitdate, l_receiptdate, l_shipinstruct, l_shipmode, l_comment Master Query -> Aggregate - Output: (sum(intermediate_column_47_0) / (sum(intermediate_column_47_1) / sum(intermediate_column_47_2))) - -> Seq Scan on pg_temp_2.pg_merge_job_0047 - Output: intermediate_column_47_0, intermediate_column_47_1, intermediate_column_47_2 + Output: (sum(intermediate_column_570007_0) / (sum(intermediate_column_570007_1) / sum(intermediate_column_570007_2))) + -> Seq Scan on pg_temp_2.pg_merge_job_570007 + Output: intermediate_column_570007_0, intermediate_column_570007_1, intermediate_column_570007_2 -- Test join EXPLAIN (COSTS FALSE) SELECT * FROM lineitem JOIN orders ON l_orderkey = o_orderkey AND l_quantity < 5 ORDER BY l_quantity DESC LIMIT 10; -Distributed Query into pg_merge_job_0048 +Distributed Query into pg_merge_job_570008 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 @@ -277,15 +279,15 @@ Distributed Query into pg_merge_job_0048 Sort Key: lineitem.l_quantity DESC -> Hash Join Hash Cond: (lineitem.l_orderkey = orders.o_orderkey) - -> Seq Scan on lineitem_102010 lineitem + -> Seq Scan on lineitem_290000 lineitem Filter: (l_quantity < '5'::numeric) -> Hash - -> Seq Scan on orders_102015 orders + -> Seq Scan on orders_290006 orders Master Query -> Limit -> Sort - Sort Key: intermediate_column_48_4 DESC - -> Seq Scan on pg_merge_job_0048 + Sort Key: intermediate_column_570008_4 DESC + -> Seq Scan on pg_merge_job_570008 -- Test insert EXPLAIN (COSTS FALSE) INSERT INTO lineitem VALUES(1,0); @@ -294,8 +296,8 @@ Distributed Query Task Count: 1 Tasks Shown: All -> Task - Node: host=localhost port=57638 dbname=regression - -> Insert on lineitem_102009 + Node: host=localhost port=57637 dbname=regression + -> Insert on lineitem_290000 -> Result -- Test update EXPLAIN (COSTS FALSE) @@ -307,12 +309,12 @@ Distributed Query Task Count: 1 Tasks Shown: All -> Task - Node: host=localhost port=57638 dbname=regression - -> Update on lineitem_102009 - -> Bitmap Heap Scan on lineitem_102009 + Node: host=localhost port=57637 dbname=regression + -> Update on lineitem_290000 + -> Bitmap Heap Scan on lineitem_290000 Recheck Cond: (l_orderkey = 1) Filter: (l_partkey = 0) - -> Bitmap Index Scan on lineitem_pkey_102009 + -> Bitmap Index Scan on lineitem_pkey_290000 Index Cond: (l_orderkey = 1) -- Test delete EXPLAIN (COSTS FALSE) @@ -323,25 +325,25 @@ Distributed Query Task Count: 1 Tasks Shown: All -> Task - Node: host=localhost port=57638 dbname=regression - -> Delete on lineitem_102009 - -> Bitmap Heap Scan on lineitem_102009 + Node: host=localhost port=57637 dbname=regression + -> Delete on lineitem_290000 + -> Bitmap Heap Scan on lineitem_290000 Recheck Cond: (l_orderkey = 1) Filter: (l_partkey = 0) - -> Bitmap Index Scan on lineitem_pkey_102009 + -> Bitmap Index Scan on lineitem_pkey_290000 Index Cond: (l_orderkey = 1) -- Test single-shard SELECT EXPLAIN (COSTS FALSE) SELECT l_quantity FROM lineitem WHERE l_orderkey = 5; -Distributed Query into pg_merge_job_0049 +Distributed Query into pg_merge_job_570009 Executor: Router Task Count: 1 Tasks Shown: All -> Task Node: host=localhost port=57637 dbname=regression - -> Bitmap Heap Scan on lineitem_102009 lineitem + -> Bitmap Heap Scan on lineitem_290000 lineitem Recheck Cond: (l_orderkey = 5) - -> Bitmap Index Scan on lineitem_pkey_102009 + -> Bitmap Index Scan on lineitem_pkey_290000 Index Cond: (l_orderkey = 5) SELECT true AS valid FROM explain_xml($$ SELECT l_quantity FROM lineitem WHERE l_orderkey = 5$$); @@ -353,41 +355,41 @@ t EXPLAIN (COSTS FALSE) CREATE TABLE explain_result AS SELECT * FROM lineitem; -Distributed Query into pg_merge_job_0052 +Distributed Query into pg_merge_job_570012 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 -> Task Node: host=localhost port=57637 dbname=regression - -> Seq Scan on lineitem_102010 lineitem + -> Seq Scan on lineitem_290000 lineitem Master Query - -> Seq Scan on pg_merge_job_0052 + -> Seq Scan on pg_merge_job_570012 -- Test all tasks output SET citus.explain_all_tasks TO on; EXPLAIN (COSTS FALSE) SELECT avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030; -Distributed Query into pg_merge_job_0053 +Distributed Query into pg_merge_job_570013 Executor: Real-Time Task Count: 3 Tasks Shown: All -> Task Node: host=localhost port=57637 dbname=regression -> Aggregate - -> Seq Scan on lineitem_102012 lineitem + -> Seq Scan on lineitem_290004 lineitem Filter: (l_orderkey > 9030) -> Task Node: host=localhost port=57638 dbname=regression -> Aggregate - -> Seq Scan on lineitem_102013 lineitem + -> Seq Scan on lineitem_290003 lineitem Filter: (l_orderkey > 9030) -> Task Node: host=localhost port=57637 dbname=regression -> Aggregate - -> Seq Scan on lineitem_102014 lineitem + -> Seq Scan on lineitem_290005 lineitem Filter: (l_orderkey > 9030) Master Query -> Aggregate - -> Seq Scan on pg_merge_job_0053 + -> Seq Scan on pg_merge_job_570013 SELECT true AS valid FROM explain_xml($$ SELECT avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030$$); t @@ -399,18 +401,18 @@ SET citus.task_executor_type TO 'task-tracker'; SET citus.explain_all_tasks TO off; EXPLAIN (COSTS FALSE) SELECT avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030; -Distributed Query into pg_merge_job_0056 +Distributed Query into pg_merge_job_570016 Executor: Task-Tracker Task Count: 3 Tasks Shown: One of 3 -> Task Node: host=localhost port=57637 dbname=regression -> Aggregate - -> Seq Scan on lineitem_102012 lineitem + -> Seq Scan on lineitem_290004 lineitem Filter: (l_orderkey > 9030) Master Query -> Aggregate - -> Seq Scan on pg_merge_job_0056 + -> Seq Scan on pg_merge_job_570016 -- Test re-partition join SET citus.large_table_shard_count TO 1; EXPLAIN (COSTS FALSE) @@ -419,7 +421,7 @@ EXPLAIN (COSTS FALSE) WHERE l_orderkey = o_orderkey AND o_custkey = c_custkey AND l_suppkey = s_suppkey; -Distributed Query into pg_merge_job_0059 +Distributed Query into pg_merge_job_570019 Executor: Task-Tracker Task Count: 1 Tasks Shown: None, not supported for re-partition queries @@ -431,7 +433,7 @@ Distributed Query into pg_merge_job_0059 Merge Task Count: 1 Master Query -> Aggregate - -> Seq Scan on pg_merge_job_0059 + -> Seq Scan on pg_merge_job_570019 EXPLAIN (COSTS FALSE, FORMAT JSON) SELECT count(*) FROM lineitem, orders, customer, supplier @@ -466,8 +468,8 @@ EXPLAIN (COSTS FALSE, FORMAT JSON) { "Node Type": "Seq Scan", "Parent Relationship": "Outer", - "Relation Name": "pg_merge_job_0062", - "Alias": "pg_merge_job_0062" + "Relation Name": "pg_merge_job_570022", + "Alias": "pg_merge_job_570022" } ] } @@ -516,8 +518,8 @@ EXPLAIN (COSTS FALSE, FORMAT XML) Seq Scan Outer - pg_merge_job_0068 - pg_merge_job_0068 + pg_merge_job_570028 + pg_merge_job_570028 @@ -555,5 +557,5 @@ EXPLAIN (COSTS FALSE, FORMAT YAML) Plans: - Node Type: "Seq Scan" Parent Relationship: "Outer" - Relation Name: "pg_merge_job_0074" - Alias: "pg_merge_job_0074" + Relation Name: "pg_merge_job_570034" + Alias: "pg_merge_job_570034" diff --git a/src/test/regress/expected/multi_explain_0.out b/src/test/regress/expected/multi_explain_0.out index 250678e0f..ff6c24944 100644 --- a/src/test/regress/expected/multi_explain_0.out +++ b/src/test/regress/expected/multi_explain_0.out @@ -1,6 +1,8 @@ -- -- MULTI_EXPLAIN -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 570000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 570000; \a\t SET citus.task_executor_type TO 'real-time'; SET citus.explain_distributed_queries TO on; @@ -30,7 +32,7 @@ $BODY$ LANGUAGE plpgsql; EXPLAIN (COSTS FALSE, FORMAT TEXT) SELECT l_quantity, count(*) count_quantity FROM lineitem GROUP BY l_quantity ORDER BY count_quantity, l_quantity; -Distributed Query into pg_merge_job_0040 +Distributed Query into pg_merge_job_570000 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 @@ -38,13 +40,13 @@ Distributed Query into pg_merge_job_0040 Node: host=localhost port=57637 dbname=regression -> HashAggregate Group Key: l_quantity - -> Seq Scan on lineitem_102010 lineitem + -> Seq Scan on lineitem_290000 lineitem Master Query -> Sort - Sort Key: (sum(((sum(intermediate_column_40_1))::bigint)))::bigint, intermediate_column_40_0 + Sort Key: (sum(((sum(intermediate_column_570000_1))::bigint)))::bigint, intermediate_column_570000_0 -> HashAggregate - Group Key: intermediate_column_40_0 - -> Seq Scan on pg_merge_job_0040 + Group Key: intermediate_column_570000_0 + -> Seq Scan on pg_merge_job_570000 -- Test JSON format EXPLAIN (COSTS FALSE, FORMAT JSON) SELECT l_quantity, count(*) count_quantity FROM lineitem @@ -69,7 +71,7 @@ EXPLAIN (COSTS FALSE, FORMAT JSON) { "Node Type": "Seq Scan", "Parent Relationship": "Outer", - "Relation Name": "lineitem_102010", + "Relation Name": "lineitem_290000", "Alias": "lineitem" } ] @@ -85,19 +87,19 @@ EXPLAIN (COSTS FALSE, FORMAT JSON) { "Plan": { "Node Type": "Sort", - "Sort Key": ["(sum(((sum(intermediate_column_41_1))::bigint)))::bigint", "intermediate_column_41_0"], + "Sort Key": ["(sum(((sum(intermediate_column_570001_1))::bigint)))::bigint", "intermediate_column_570001_0"], "Plans": [ { "Node Type": "Aggregate", "Strategy": "Hashed", "Parent Relationship": "Outer", - "Group Key": ["intermediate_column_41_0"], + "Group Key": ["intermediate_column_570001_0"], "Plans": [ { "Node Type": "Seq Scan", "Parent Relationship": "Outer", - "Relation Name": "pg_merge_job_0041", - "Alias": "pg_merge_job_0041" + "Relation Name": "pg_merge_job_570001", + "Alias": "pg_merge_job_570001" } ] } @@ -138,7 +140,7 @@ EXPLAIN (COSTS FALSE, FORMAT XML) Seq Scan Outer - lineitem_102010 + lineitem_290000 lineitem @@ -154,8 +156,8 @@ EXPLAIN (COSTS FALSE, FORMAT XML) Sort - (sum(((sum(intermediate_column_43_1))::bigint)))::bigint - intermediate_column_43_0 + (sum(((sum(intermediate_column_570003_1))::bigint)))::bigint + intermediate_column_570003_0 @@ -163,14 +165,14 @@ EXPLAIN (COSTS FALSE, FORMAT XML) Hashed Outer - intermediate_column_43_0 + intermediate_column_570003_0 Seq Scan Outer - pg_merge_job_0043 - pg_merge_job_0043 + pg_merge_job_570003 + pg_merge_job_570003 @@ -204,31 +206,31 @@ EXPLAIN (COSTS FALSE, FORMAT YAML) Plans: - Node Type: "Seq Scan" Parent Relationship: "Outer" - Relation Name: "lineitem_102010" + Relation Name: "lineitem_290000" Alias: "lineitem" Master Query: - Plan: Node Type: "Sort" Sort Key: - - "(sum(((sum(intermediate_column_45_1))::bigint)))::bigint" - - "intermediate_column_45_0" + - "(sum(((sum(intermediate_column_570005_1))::bigint)))::bigint" + - "intermediate_column_570005_0" Plans: - Node Type: "Aggregate" Strategy: "Hashed" Parent Relationship: "Outer" Group Key: - - "intermediate_column_45_0" + - "intermediate_column_570005_0" Plans: - Node Type: "Seq Scan" Parent Relationship: "Outer" - Relation Name: "pg_merge_job_0045" - Alias: "pg_merge_job_0045" + Relation Name: "pg_merge_job_570005" + Alias: "pg_merge_job_570005" -- Test Text format EXPLAIN (COSTS FALSE, FORMAT TEXT) SELECT l_quantity, count(*) count_quantity FROM lineitem GROUP BY l_quantity ORDER BY count_quantity, l_quantity; -Distributed Query into pg_merge_job_0046 +Distributed Query into pg_merge_job_570006 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 @@ -236,17 +238,17 @@ Distributed Query into pg_merge_job_0046 Node: host=localhost port=57637 dbname=regression -> HashAggregate Group Key: l_quantity - -> Seq Scan on lineitem_102010 lineitem + -> Seq Scan on lineitem_290000 lineitem Master Query -> Sort - Sort Key: (sum(((sum(intermediate_column_46_1))::bigint)))::bigint, intermediate_column_46_0 + Sort Key: (sum(((sum(intermediate_column_570006_1))::bigint)))::bigint, intermediate_column_570006_0 -> HashAggregate - Group Key: intermediate_column_46_0 - -> Seq Scan on pg_merge_job_0046 + Group Key: intermediate_column_570006_0 + -> Seq Scan on pg_merge_job_570006 -- Test verbose EXPLAIN (COSTS FALSE, VERBOSE TRUE) SELECT sum(l_quantity) / avg(l_quantity) FROM lineitem; -Distributed Query into pg_merge_job_0047 +Distributed Query into pg_merge_job_570007 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 @@ -254,19 +256,19 @@ Distributed Query into pg_merge_job_0047 Node: host=localhost port=57637 dbname=regression -> Aggregate Output: sum(l_quantity), sum(l_quantity), count(l_quantity) - -> Seq Scan on public.lineitem_102010 lineitem + -> Seq Scan on public.lineitem_290000 lineitem Output: l_orderkey, l_partkey, l_suppkey, l_linenumber, l_quantity, l_extendedprice, l_discount, l_tax, l_returnflag, l_linestatus, l_shipdate, l_commitdate, l_receiptdate, l_shipinstruct, l_shipmode, l_comment Master Query -> Aggregate - Output: (sum(intermediate_column_47_0) / (sum(intermediate_column_47_1) / sum(intermediate_column_47_2))) - -> Seq Scan on pg_temp_2.pg_merge_job_0047 - Output: intermediate_column_47_0, intermediate_column_47_1, intermediate_column_47_2 + Output: (sum(intermediate_column_570007_0) / (sum(intermediate_column_570007_1) / sum(intermediate_column_570007_2))) + -> Seq Scan on pg_temp_2.pg_merge_job_570007 + Output: intermediate_column_570007_0, intermediate_column_570007_1, intermediate_column_570007_2 -- Test join EXPLAIN (COSTS FALSE) SELECT * FROM lineitem JOIN orders ON l_orderkey = o_orderkey AND l_quantity < 5 ORDER BY l_quantity DESC LIMIT 10; -Distributed Query into pg_merge_job_0048 +Distributed Query into pg_merge_job_570008 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 @@ -277,15 +279,15 @@ Distributed Query into pg_merge_job_0048 Sort Key: lineitem.l_quantity -> Hash Join Hash Cond: (lineitem.l_orderkey = orders.o_orderkey) - -> Seq Scan on lineitem_102010 lineitem + -> Seq Scan on lineitem_290000 lineitem Filter: (l_quantity < 5::numeric) -> Hash - -> Seq Scan on orders_102015 orders + -> Seq Scan on orders_290006 orders Master Query -> Limit -> Sort - Sort Key: intermediate_column_48_4 - -> Seq Scan on pg_merge_job_0048 + Sort Key: intermediate_column_570008_4 + -> Seq Scan on pg_merge_job_570008 -- Test insert EXPLAIN (COSTS FALSE) INSERT INTO lineitem VALUES(1,0); @@ -294,8 +296,8 @@ Distributed Query Task Count: 1 Tasks Shown: All -> Task - Node: host=localhost port=57638 dbname=regression - -> Insert on lineitem_102009 + Node: host=localhost port=57637 dbname=regression + -> Insert on lineitem_290000 -> Result -- Test update EXPLAIN (COSTS FALSE) @@ -307,12 +309,12 @@ Distributed Query Task Count: 1 Tasks Shown: All -> Task - Node: host=localhost port=57638 dbname=regression - -> Update on lineitem_102009 - -> Bitmap Heap Scan on lineitem_102009 + Node: host=localhost port=57637 dbname=regression + -> Update on lineitem_290000 + -> Bitmap Heap Scan on lineitem_290000 Recheck Cond: (l_orderkey = 1) Filter: (l_partkey = 0) - -> Bitmap Index Scan on lineitem_pkey_102009 + -> Bitmap Index Scan on lineitem_pkey_290000 Index Cond: (l_orderkey = 1) -- Test delete EXPLAIN (COSTS FALSE) @@ -323,25 +325,25 @@ Distributed Query Task Count: 1 Tasks Shown: All -> Task - Node: host=localhost port=57638 dbname=regression - -> Delete on lineitem_102009 - -> Bitmap Heap Scan on lineitem_102009 + Node: host=localhost port=57637 dbname=regression + -> Delete on lineitem_290000 + -> Bitmap Heap Scan on lineitem_290000 Recheck Cond: (l_orderkey = 1) Filter: (l_partkey = 0) - -> Bitmap Index Scan on lineitem_pkey_102009 + -> Bitmap Index Scan on lineitem_pkey_290000 Index Cond: (l_orderkey = 1) -- Test single-shard SELECT EXPLAIN (COSTS FALSE) SELECT l_quantity FROM lineitem WHERE l_orderkey = 5; -Distributed Query into pg_merge_job_0049 +Distributed Query into pg_merge_job_570009 Executor: Router Task Count: 1 Tasks Shown: All -> Task Node: host=localhost port=57637 dbname=regression - -> Bitmap Heap Scan on lineitem_102009 lineitem + -> Bitmap Heap Scan on lineitem_290000 lineitem Recheck Cond: (l_orderkey = 5) - -> Bitmap Index Scan on lineitem_pkey_102009 + -> Bitmap Index Scan on lineitem_pkey_290000 Index Cond: (l_orderkey = 5) SELECT true AS valid FROM explain_xml($$ SELECT l_quantity FROM lineitem WHERE l_orderkey = 5$$); @@ -353,41 +355,41 @@ t EXPLAIN (COSTS FALSE) CREATE TABLE explain_result AS SELECT * FROM lineitem; -Distributed Query into pg_merge_job_0052 +Distributed Query into pg_merge_job_570012 Executor: Real-Time Task Count: 6 Tasks Shown: One of 6 -> Task Node: host=localhost port=57637 dbname=regression - -> Seq Scan on lineitem_102010 lineitem + -> Seq Scan on lineitem_290000 lineitem Master Query - -> Seq Scan on pg_merge_job_0052 + -> Seq Scan on pg_merge_job_570012 -- Test all tasks output SET citus.explain_all_tasks TO on; EXPLAIN (COSTS FALSE) SELECT avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030; -Distributed Query into pg_merge_job_0053 +Distributed Query into pg_merge_job_570013 Executor: Real-Time Task Count: 3 Tasks Shown: All -> Task Node: host=localhost port=57637 dbname=regression -> Aggregate - -> Seq Scan on lineitem_102012 lineitem + -> Seq Scan on lineitem_290004 lineitem Filter: (l_orderkey > 9030) -> Task Node: host=localhost port=57638 dbname=regression -> Aggregate - -> Seq Scan on lineitem_102013 lineitem + -> Seq Scan on lineitem_290003 lineitem Filter: (l_orderkey > 9030) -> Task Node: host=localhost port=57637 dbname=regression -> Aggregate - -> Seq Scan on lineitem_102014 lineitem + -> Seq Scan on lineitem_290005 lineitem Filter: (l_orderkey > 9030) Master Query -> Aggregate - -> Seq Scan on pg_merge_job_0053 + -> Seq Scan on pg_merge_job_570013 SELECT true AS valid FROM explain_xml($$ SELECT avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030$$); t @@ -399,18 +401,18 @@ SET citus.task_executor_type TO 'task-tracker'; SET citus.explain_all_tasks TO off; EXPLAIN (COSTS FALSE) SELECT avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030; -Distributed Query into pg_merge_job_0056 +Distributed Query into pg_merge_job_570016 Executor: Task-Tracker Task Count: 3 Tasks Shown: One of 3 -> Task Node: host=localhost port=57637 dbname=regression -> Aggregate - -> Seq Scan on lineitem_102012 lineitem + -> Seq Scan on lineitem_290004 lineitem Filter: (l_orderkey > 9030) Master Query -> Aggregate - -> Seq Scan on pg_merge_job_0056 + -> Seq Scan on pg_merge_job_570016 -- Test re-partition join SET citus.large_table_shard_count TO 1; EXPLAIN (COSTS FALSE) @@ -419,7 +421,7 @@ EXPLAIN (COSTS FALSE) WHERE l_orderkey = o_orderkey AND o_custkey = c_custkey AND l_suppkey = s_suppkey; -Distributed Query into pg_merge_job_0059 +Distributed Query into pg_merge_job_570019 Executor: Task-Tracker Task Count: 1 Tasks Shown: None, not supported for re-partition queries @@ -431,7 +433,7 @@ Distributed Query into pg_merge_job_0059 Merge Task Count: 1 Master Query -> Aggregate - -> Seq Scan on pg_merge_job_0059 + -> Seq Scan on pg_merge_job_570019 EXPLAIN (COSTS FALSE, FORMAT JSON) SELECT count(*) FROM lineitem, orders, customer, supplier @@ -466,8 +468,8 @@ EXPLAIN (COSTS FALSE, FORMAT JSON) { "Node Type": "Seq Scan", "Parent Relationship": "Outer", - "Relation Name": "pg_merge_job_0062", - "Alias": "pg_merge_job_0062" + "Relation Name": "pg_merge_job_570022", + "Alias": "pg_merge_job_570022" } ] } @@ -516,8 +518,8 @@ EXPLAIN (COSTS FALSE, FORMAT XML) Seq Scan Outer - pg_merge_job_0068 - pg_merge_job_0068 + pg_merge_job_570028 + pg_merge_job_570028 @@ -555,5 +557,5 @@ EXPLAIN (COSTS FALSE, FORMAT YAML) Plans: - Node Type: "Seq Scan" Parent Relationship: "Outer" - Relation Name: "pg_merge_job_0074" - Alias: "pg_merge_job_0074" + Relation Name: "pg_merge_job_570034" + Alias: "pg_merge_job_570034" diff --git a/src/test/regress/expected/multi_extension.out b/src/test/regress/expected/multi_extension.out index 4983208f6..3228d2e30 100644 --- a/src/test/regress/expected/multi_extension.out +++ b/src/test/regress/expected/multi_extension.out @@ -5,6 +5,8 @@ -- -- It'd be nice to script generation of this file, but alas, that's -- not done yet. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 580000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 580000; -- DROP EXTENSION pre-created by the regression suite DROP EXTENSION citus; \c diff --git a/src/test/regress/expected/multi_fdw_create_table.out b/src/test/regress/expected/multi_fdw_create_table.out index 6557e318a..8facf2d81 100644 --- a/src/test/regress/expected/multi_fdw_create_table.out +++ b/src/test/regress/expected/multi_fdw_create_table.out @@ -1,6 +1,8 @@ -- -- MULTI_FDW_CREATE_TABLE -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 590000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 590000; -- Create new table definitions for use in testing in distributed foreign data -- wrapper functionality. SELECT fdwname FROM pg_foreign_data_wrapper; diff --git a/src/test/regress/expected/multi_fdw_master_protocol.out b/src/test/regress/expected/multi_fdw_master_protocol.out index bffb6d356..8fb73ac13 100644 --- a/src/test/regress/expected/multi_fdw_master_protocol.out +++ b/src/test/regress/expected/multi_fdw_master_protocol.out @@ -2,6 +2,8 @@ -- MULTI_FDW_MASTER_PROTOCOL -- -- Tests that check the metadata returned by the master node. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 600000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 600000; SELECT part_storage_type, part_key, part_replica_count, part_max_size, part_placement_policy FROM master_get_table_metadata('lineitem'); part_storage_type | part_key | part_replica_count | part_max_size | part_placement_policy @@ -20,7 +22,7 @@ SELECT * FROM master_get_table_ddl_events('lineitem'); SELECT * FROM master_get_new_shardid(); master_get_new_shardid ------------------------ - 102008 + 600000 (1 row) SELECT node_name FROM master_get_local_first_candidate_nodes(); diff --git a/src/test/regress/expected/multi_generate_ddl_commands.out b/src/test/regress/expected/multi_generate_ddl_commands.out index ff856b423..bb929e533 100644 --- a/src/test/regress/expected/multi_generate_ddl_commands.out +++ b/src/test/regress/expected/multi_generate_ddl_commands.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 610000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 610000; -- =================================================================== -- create test functions -- =================================================================== diff --git a/src/test/regress/expected/multi_hash_pruning.out b/src/test/regress/expected/multi_hash_pruning.out index 84a977fce..c7dc9e30b 100644 --- a/src/test/regress/expected/multi_hash_pruning.out +++ b/src/test/regress/expected/multi_hash_pruning.out @@ -2,6 +2,8 @@ -- MULTI_HASH_PRUNING -- -- Tests for shard and join pruning logic on hash partitioned tables. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 630000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 630000; -- Create a table partitioned on integer column and update partition type to -- hash. Then stage data to this table and update shard min max values with -- hashed ones. Hash value of 1, 2, 3 and 4 are consecutively -1905060026, @@ -47,9 +49,9 @@ SELECT count(*) FROM orders_hash_partitioned; SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 DEBUG: Plan is router executable count ------- @@ -58,9 +60,9 @@ DEBUG: Plan is router executable SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 2; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 102033 -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 +DEBUG: predicate pruning for shardId 630000 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 DEBUG: Plan is router executable count ------- @@ -69,9 +71,9 @@ DEBUG: Plan is router executable SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 3; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 102033 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630000 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 DEBUG: Plan is router executable count ------- @@ -80,9 +82,9 @@ DEBUG: Plan is router executable SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 4; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 102033 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630000 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 DEBUG: Plan is router executable count ------- @@ -92,9 +94,9 @@ DEBUG: Plan is router executable SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 1 AND o_clerk = 'aaa'; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 DEBUG: Plan is router executable count ------- @@ -103,9 +105,9 @@ DEBUG: Plan is router executable SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = abs(-1); DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 DEBUG: Plan is router executable count ------- @@ -120,36 +122,36 @@ SELECT count(*) FROM orders_hash_partitioned; (1 row) SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 1; -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 count ------- 0 (1 row) SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 2; -DEBUG: predicate pruning for shardId 102033 -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 +DEBUG: predicate pruning for shardId 630000 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 count ------- 0 (1 row) SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 3; -DEBUG: predicate pruning for shardId 102033 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630000 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 count ------- 0 (1 row) SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 4; -DEBUG: predicate pruning for shardId 102033 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630000 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 count ------- 0 @@ -157,18 +159,18 @@ DEBUG: predicate pruning for shardId 102036 SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 1 AND o_clerk = 'aaa'; -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 count ------- 0 (1 row) SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = abs(-1); -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 count ------- 0 @@ -176,9 +178,9 @@ DEBUG: predicate pruning for shardId 102036 SET citus.task_executor_type TO :actual_task_executor; SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey is NULL; -DEBUG: predicate pruning for shardId 102033 -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630000 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630003 count ------- 0 @@ -198,8 +200,8 @@ SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey > 2; SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 1 OR o_orderkey = 2; -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 count ------- 0 @@ -214,8 +216,8 @@ SELECT count(*) FROM orders_hash_partitioned SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 1 OR (o_orderkey = 3 AND o_clerk = 'aaa'); -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 count ------- 0 @@ -223,8 +225,8 @@ DEBUG: predicate pruning for shardId 102036 SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = 1 OR o_orderkey is NULL; -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630003 count ------- 0 @@ -232,9 +234,9 @@ DEBUG: predicate pruning for shardId 102036 SELECT count(*) FROM (SELECT o_orderkey FROM orders_hash_partitioned WHERE o_orderkey = 1) AS orderkeys; -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 count ------- 0 @@ -285,9 +287,9 @@ SELECT count(*) FROM orders_hash_partitioned SELECT count(*) FROM orders_hash_partitioned WHERE o_orderkey = random() AND o_orderkey = 1; -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 count ------- 0 @@ -319,12 +321,12 @@ SELECT count(*) WHERE orders1.o_orderkey = orders2.o_orderkey AND orders1.o_orderkey = 1 AND orders2.o_orderkey is NULL; -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102035 -DEBUG: predicate pruning for shardId 102036 -DEBUG: predicate pruning for shardId 102033 -DEBUG: predicate pruning for shardId 102034 -DEBUG: predicate pruning for shardId 102036 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630002 +DEBUG: predicate pruning for shardId 630003 +DEBUG: predicate pruning for shardId 630000 +DEBUG: predicate pruning for shardId 630001 +DEBUG: predicate pruning for shardId 630003 DEBUG: join prunable for intervals [-2147483648,-1073741825] and [0,1073741823] count ------- diff --git a/src/test/regress/expected/multi_index_statements.out b/src/test/regress/expected/multi_index_statements.out index 2fd7a89e6..b8258596f 100644 --- a/src/test/regress/expected/multi_index_statements.out +++ b/src/test/regress/expected/multi_index_statements.out @@ -3,6 +3,8 @@ -- -- Check that we can run CREATE INDEX and DROP INDEX statements on distributed -- tables. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 640000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 640000; -- -- CREATE TEST TABLES -- @@ -137,11 +139,11 @@ ERROR: creating unique indexes on append-partitioned tables is currently unsupp CREATE INDEX lineitem_orderkey_index ON lineitem (l_orderkey); ERROR: relation "lineitem_orderkey_index" already exists CREATE INDEX try_index ON lineitem USING gist (l_orderkey); -WARNING: Bad result from localhost:57637 +WARNING: Bad result from localhost:57638 DETAIL: Remote message: data type bigint has no default operator class for access method "gist" ERROR: could not execute DDL command on worker node shards CREATE INDEX try_index ON lineitem (non_existent_column); -WARNING: Bad result from localhost:57637 +WARNING: Bad result from localhost:57638 DETAIL: Remote message: column "non_existent_column" does not exist ERROR: could not execute DDL command on worker node shards CREATE INDEX ON lineitem (l_orderkey); diff --git a/src/test/regress/expected/multi_join_order_additional.out b/src/test/regress/expected/multi_join_order_additional.out index f4380134d..20698676a 100644 --- a/src/test/regress/expected/multi_join_order_additional.out +++ b/src/test/regress/expected/multi_join_order_additional.out @@ -1,6 +1,8 @@ -- -- MULTI_JOIN_ORDER_ADDITIONAL -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 650000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 650000; -- Set configuration to print table join order and pruned shards SET citus.explain_distributed_queries TO off; SET citus.log_multi_join_order TO TRUE; @@ -40,8 +42,8 @@ SELECT master_create_worker_shards('lineitem_hash', 2, 1); (1 row) CREATE INDEX lineitem_hash_time_index ON lineitem_hash (l_shipdate); -DEBUG: applied command on shard 102037 on node localhost:57637 -DEBUG: applied command on shard 102038 on node localhost:57638 +DEBUG: applied command on shard 650000 on node localhost:57637 +DEBUG: applied command on shard 650001 on node localhost:57638 DEBUG: building index "lineitem_hash_time_index" on table "lineitem_hash" CREATE TABLE orders_hash ( o_orderkey bigint not null, diff --git a/src/test/regress/expected/multi_join_order_tpch_large.out b/src/test/regress/expected/multi_join_order_tpch_large.out index 04759db3b..6532d9ea1 100644 --- a/src/test/regress/expected/multi_join_order_tpch_large.out +++ b/src/test/regress/expected/multi_join_order_tpch_large.out @@ -1,6 +1,8 @@ -- -- MULTI_JOIN_ORDER_TPCH_LARGE -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 660000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 660000; -- Enable configuration to print table join order SET citus.explain_distributed_queries TO off; SET citus.log_multi_join_order TO TRUE; diff --git a/src/test/regress/expected/multi_join_pruning.out b/src/test/regress/expected/multi_join_pruning.out index 2b2968321..fd77d607b 100644 --- a/src/test/regress/expected/multi_join_pruning.out +++ b/src/test/regress/expected/multi_join_pruning.out @@ -1,6 +1,7 @@ -- -- MULTI_JOIN_PRUNING -- +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 680000; -- Check that join-pruning works for joins between two large relations. For now -- we only check for join-pruning between locally partitioned relations. In the -- future we want to check for pruning between re-partitioned relations as well. @@ -23,9 +24,9 @@ DEBUG: join prunable for intervals [13921,14947] and [1,5986] SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem, orders WHERE l_orderkey = o_orderkey AND l_orderkey > 9030; -DEBUG: predicate pruning for shardId 102009 -DEBUG: predicate pruning for shardId 102010 -DEBUG: predicate pruning for shardId 102011 +DEBUG: predicate pruning for shardId 290000 +DEBUG: predicate pruning for shardId 290001 +DEBUG: predicate pruning for shardId 290002 DEBUG: join prunable for intervals [8997,11554] and [1,5986] DEBUG: join prunable for intervals [11554,13920] and [1,5986] DEBUG: join prunable for intervals [13921,14947] and [1,5986] @@ -38,12 +39,12 @@ DEBUG: join prunable for intervals [13921,14947] and [1,5986] -- works as expected in this case. SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem, orders WHERE l_orderkey = o_orderkey AND l_orderkey > 20000; -DEBUG: predicate pruning for shardId 102009 -DEBUG: predicate pruning for shardId 102010 -DEBUG: predicate pruning for shardId 102011 -DEBUG: predicate pruning for shardId 102012 -DEBUG: predicate pruning for shardId 102013 -DEBUG: predicate pruning for shardId 102014 +DEBUG: predicate pruning for shardId 290000 +DEBUG: predicate pruning for shardId 290001 +DEBUG: predicate pruning for shardId 290002 +DEBUG: predicate pruning for shardId 290003 +DEBUG: predicate pruning for shardId 290004 +DEBUG: predicate pruning for shardId 290005 sum | avg -----+----- | @@ -54,10 +55,10 @@ DEBUG: predicate pruning for shardId 102014 -- out all the shards, and leave us with an empty task list. SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem, orders WHERE l_orderkey = o_orderkey AND l_orderkey > 6000 AND o_orderkey < 6000; -DEBUG: predicate pruning for shardId 102009 -DEBUG: predicate pruning for shardId 102010 -DEBUG: predicate pruning for shardId 102011 -DEBUG: predicate pruning for shardId 102016 +DEBUG: predicate pruning for shardId 290000 +DEBUG: predicate pruning for shardId 290001 +DEBUG: predicate pruning for shardId 290002 +DEBUG: predicate pruning for shardId 290007 DEBUG: join prunable for intervals [8997,11554] and [1,5986] DEBUG: join prunable for intervals [11554,13920] and [1,5986] DEBUG: join prunable for intervals [13921,14947] and [1,5986] diff --git a/src/test/regress/expected/multi_large_table_join_planning.out b/src/test/regress/expected/multi_large_table_join_planning.out index ee2491426..70dfe85a0 100644 --- a/src/test/regress/expected/multi_large_table_join_planning.out +++ b/src/test/regress/expected/multi_large_table_join_planning.out @@ -5,6 +5,8 @@ -- transaction block here so that we don't emit debug messages with changing -- transaction ids in them. Also, we set the executor type to task tracker -- executor here, as we cannot run repartition jobs with real time executor. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 690000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 690000; BEGIN; SET client_min_messages TO DEBUG4; DEBUG: CommitTransactionCommand @@ -52,29 +54,29 @@ DEBUG: join prunable for intervals [8997,11554] and [1,5986] DEBUG: join prunable for intervals [11554,13920] and [1,5986] DEBUG: join prunable for intervals [13921,14947] and [1,5986] DEBUG: generated sql query for job 1250 and task 3 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102009 lineitem JOIN orders_102015 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290000 lineitem JOIN orders_290006 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" DEBUG: generated sql query for job 1250 and task 6 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102010 lineitem JOIN orders_102015 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290001 lineitem JOIN orders_290006 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" DEBUG: generated sql query for job 1250 and task 9 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102011 lineitem JOIN orders_102015 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290002 lineitem JOIN orders_290006 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" DEBUG: generated sql query for job 1250 and task 12 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102012 lineitem JOIN orders_102016 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290003 lineitem JOIN orders_290007 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" DEBUG: generated sql query for job 1250 and task 15 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102013 lineitem JOIN orders_102016 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290004 lineitem JOIN orders_290007 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" DEBUG: generated sql query for job 1250 and task 18 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102014 lineitem JOIN orders_102016 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" -DEBUG: assigned task 6 to node localhost:57637 -DEBUG: assigned task 3 to node localhost:57638 -DEBUG: assigned task 12 to node localhost:57637 -DEBUG: assigned task 9 to node localhost:57638 -DEBUG: assigned task 18 to node localhost:57637 -DEBUG: assigned task 15 to node localhost:57638 +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290005 lineitem JOIN orders_290007 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > '10'::numeric)" +DEBUG: assigned task 3 to node localhost:57637 +DEBUG: assigned task 6 to node localhost:57638 +DEBUG: assigned task 9 to node localhost:57637 +DEBUG: assigned task 12 to node localhost:57638 +DEBUG: assigned task 15 to node localhost:57637 +DEBUG: assigned task 18 to node localhost:57638 DEBUG: join prunable for intervals [1,1000] and [6001,7000] DEBUG: join prunable for intervals [6001,7000] and [1,1000] DEBUG: generated sql query for job 1251 and task 3 -DETAIL: query string: "SELECT "pg_merge_job_1250.task_000019".intermediate_column_1250_0, "pg_merge_job_1250.task_000019".intermediate_column_1250_1, "pg_merge_job_1250.task_000019".intermediate_column_1250_2, "pg_merge_job_1250.task_000019".intermediate_column_1250_3, "pg_merge_job_1250.task_000019".intermediate_column_1250_4 FROM (pg_merge_job_1250.task_000019 "pg_merge_job_1250.task_000019" JOIN part_102019 part ON (("pg_merge_job_1250.task_000019".intermediate_column_1250_0 = part.p_partkey))) WHERE (part.p_size > 8)" +DETAIL: query string: "SELECT "pg_merge_job_1250.task_000019".intermediate_column_1250_0, "pg_merge_job_1250.task_000019".intermediate_column_1250_1, "pg_merge_job_1250.task_000019".intermediate_column_1250_2, "pg_merge_job_1250.task_000019".intermediate_column_1250_3, "pg_merge_job_1250.task_000019".intermediate_column_1250_4 FROM (pg_merge_job_1250.task_000019 "pg_merge_job_1250.task_000019" JOIN part_290010 part ON (("pg_merge_job_1250.task_000019".intermediate_column_1250_0 = part.p_partkey))) WHERE (part.p_size > 8)" DEBUG: generated sql query for job 1251 and task 6 -DETAIL: query string: "SELECT "pg_merge_job_1250.task_000026".intermediate_column_1250_0, "pg_merge_job_1250.task_000026".intermediate_column_1250_1, "pg_merge_job_1250.task_000026".intermediate_column_1250_2, "pg_merge_job_1250.task_000026".intermediate_column_1250_3, "pg_merge_job_1250.task_000026".intermediate_column_1250_4 FROM (pg_merge_job_1250.task_000026 "pg_merge_job_1250.task_000026" JOIN part_102044 part ON (("pg_merge_job_1250.task_000026".intermediate_column_1250_0 = part.p_partkey))) WHERE (part.p_size > 8)" +DETAIL: query string: "SELECT "pg_merge_job_1250.task_000026".intermediate_column_1250_0, "pg_merge_job_1250.task_000026".intermediate_column_1250_1, "pg_merge_job_1250.task_000026".intermediate_column_1250_2, "pg_merge_job_1250.task_000026".intermediate_column_1250_3, "pg_merge_job_1250.task_000026".intermediate_column_1250_4 FROM (pg_merge_job_1250.task_000026 "pg_merge_job_1250.task_000026" JOIN part_280002 part ON (("pg_merge_job_1250.task_000026".intermediate_column_1250_0 = part.p_partkey))) WHERE (part.p_size > 8)" DEBUG: pruning merge fetch taskId 1 DETAIL: Creating dependency on merge taskId 19 DEBUG: pruning merge fetch taskId 4 @@ -88,11 +90,11 @@ DEBUG: join prunable for intervals [1001,2000] and [6001,7000] DEBUG: join prunable for intervals [6001,7000] and [1,1000] DEBUG: join prunable for intervals [6001,7000] and [1001,2000] DEBUG: generated sql query for job 1252 and task 3 -DETAIL: query string: "SELECT "pg_merge_job_1251.task_000007".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000007 "pg_merge_job_1251.task_000007" JOIN customer_102017 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000007".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000007".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000007".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1, "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 LIMIT '30'::bigint" +DETAIL: query string: "SELECT "pg_merge_job_1251.task_000007".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000007 "pg_merge_job_1251.task_000007" JOIN customer_290008 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000007".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000007".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000007".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1, "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 LIMIT '30'::bigint" DEBUG: generated sql query for job 1252 and task 6 -DETAIL: query string: "SELECT "pg_merge_job_1251.task_000010".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000010 "pg_merge_job_1251.task_000010" JOIN customer_102043 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000010".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000010".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000010".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1, "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 LIMIT '30'::bigint" +DETAIL: query string: "SELECT "pg_merge_job_1251.task_000010".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000010 "pg_merge_job_1251.task_000010" JOIN customer_280001 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000010".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000010".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000010".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1, "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 LIMIT '30'::bigint" DEBUG: generated sql query for job 1252 and task 9 -DETAIL: query string: "SELECT "pg_merge_job_1251.task_000013".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000013 "pg_merge_job_1251.task_000013" JOIN customer_102042 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000013".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000013".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000013".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1, "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 LIMIT '30'::bigint" +DETAIL: query string: "SELECT "pg_merge_job_1251.task_000013".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000013 "pg_merge_job_1251.task_000013" JOIN customer_280000 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000013".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000013".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000013".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1, "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 LIMIT '30'::bigint" DEBUG: pruning merge fetch taskId 1 DETAIL: Creating dependency on merge taskId 7 DEBUG: pruning merge fetch taskId 4 @@ -100,8 +102,8 @@ DETAIL: Creating dependency on merge taskId 10 DEBUG: pruning merge fetch taskId 7 DETAIL: Creating dependency on merge taskId 13 DEBUG: assigned task 9 to node localhost:57637 -DEBUG: assigned task 3 to node localhost:57638 -DEBUG: assigned task 6 to node localhost:57637 +DEBUG: assigned task 6 to node localhost:57638 +DEBUG: assigned task 3 to node localhost:57637 DEBUG: completed cleanup query for job 1252 on node "localhost:57638" DEBUG: completed cleanup query for job 1252 on node "localhost:57637" DEBUG: completed cleanup query for job 1251 on node "localhost:57638" @@ -156,29 +158,29 @@ ORDER BY l_partkey, o_orderkey; DEBUG: StartTransactionCommand DEBUG: generated sql query for job 1253 and task 2 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102009 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290000 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 4 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102010 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290001 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 6 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102011 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290002 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 8 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102012 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290003 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 10 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102013 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290004 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 12 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102014 lineitem WHERE (l_quantity < 5.0)" -DEBUG: assigned task 4 to node localhost:57637 -DEBUG: assigned task 2 to node localhost:57638 -DEBUG: assigned task 8 to node localhost:57637 -DEBUG: assigned task 6 to node localhost:57638 -DEBUG: assigned task 12 to node localhost:57637 -DEBUG: assigned task 10 to node localhost:57638 +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290005 lineitem WHERE (l_quantity < 5.0)" +DEBUG: assigned task 2 to node localhost:57637 +DEBUG: assigned task 4 to node localhost:57638 +DEBUG: assigned task 6 to node localhost:57637 +DEBUG: assigned task 8 to node localhost:57638 +DEBUG: assigned task 10 to node localhost:57637 +DEBUG: assigned task 12 to node localhost:57638 DEBUG: generated sql query for job 1254 and task 2 -DETAIL: query string: "SELECT o_orderkey, o_shippriority FROM orders_102015 orders WHERE (o_totalprice <> 4.0)" +DETAIL: query string: "SELECT o_orderkey, o_shippriority FROM orders_290006 orders WHERE (o_totalprice <> 4.0)" DEBUG: generated sql query for job 1254 and task 4 -DETAIL: query string: "SELECT o_orderkey, o_shippriority FROM orders_102016 orders WHERE (o_totalprice <> 4.0)" -DEBUG: assigned task 4 to node localhost:57637 -DEBUG: assigned task 2 to node localhost:57638 +DETAIL: query string: "SELECT o_orderkey, o_shippriority FROM orders_290007 orders WHERE (o_totalprice <> 4.0)" +DEBUG: assigned task 2 to node localhost:57637 +DEBUG: assigned task 4 to node localhost:57638 DEBUG: join prunable for task partitionId 0 and 1 DEBUG: join prunable for task partitionId 0 and 2 DEBUG: join prunable for task partitionId 0 and 3 diff --git a/src/test/regress/expected/multi_large_table_join_planning_0.out b/src/test/regress/expected/multi_large_table_join_planning_0.out index 7ea71fb38..65c0b8ca0 100644 --- a/src/test/regress/expected/multi_large_table_join_planning_0.out +++ b/src/test/regress/expected/multi_large_table_join_planning_0.out @@ -5,6 +5,8 @@ -- transaction block here so that we don't emit debug messages with changing -- transaction ids in them. Also, we set the executor type to task tracker -- executor here, as we cannot run repartition jobs with real time executor. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 690000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 690000; BEGIN; SET client_min_messages TO DEBUG4; DEBUG: CommitTransactionCommand @@ -52,29 +54,29 @@ DEBUG: join prunable for intervals [8997,11554] and [1,5986] DEBUG: join prunable for intervals [11554,13920] and [1,5986] DEBUG: join prunable for intervals [13921,14947] and [1,5986] DEBUG: generated sql query for job 1250 and task 3 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102009 lineitem JOIN orders_102015 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290000 lineitem JOIN orders_290006 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" DEBUG: generated sql query for job 1250 and task 6 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102010 lineitem JOIN orders_102015 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290001 lineitem JOIN orders_290006 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" DEBUG: generated sql query for job 1250 and task 9 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102011 lineitem JOIN orders_102015 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290002 lineitem JOIN orders_290006 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" DEBUG: generated sql query for job 1250 and task 12 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102012 lineitem JOIN orders_102016 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290003 lineitem JOIN orders_290007 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" DEBUG: generated sql query for job 1250 and task 15 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102013 lineitem JOIN orders_102016 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290004 lineitem JOIN orders_290007 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" DEBUG: generated sql query for job 1250 and task 18 -DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_102014 lineitem JOIN orders_102016 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" -DEBUG: assigned task 6 to node localhost:57637 -DEBUG: assigned task 3 to node localhost:57638 -DEBUG: assigned task 12 to node localhost:57637 -DEBUG: assigned task 9 to node localhost:57638 -DEBUG: assigned task 18 to node localhost:57637 -DEBUG: assigned task 15 to node localhost:57638 +DETAIL: query string: "SELECT lineitem.l_partkey, orders.o_orderkey, lineitem.l_quantity, lineitem.l_extendedprice, orders.o_custkey FROM (lineitem_290005 lineitem JOIN orders_290007 orders ON ((lineitem.l_orderkey = orders.o_orderkey))) WHERE (orders.o_totalprice > 10::numeric)" +DEBUG: assigned task 3 to node localhost:57637 +DEBUG: assigned task 6 to node localhost:57638 +DEBUG: assigned task 9 to node localhost:57637 +DEBUG: assigned task 12 to node localhost:57638 +DEBUG: assigned task 15 to node localhost:57637 +DEBUG: assigned task 18 to node localhost:57638 DEBUG: join prunable for intervals [1,1000] and [6001,7000] DEBUG: join prunable for intervals [6001,7000] and [1,1000] DEBUG: generated sql query for job 1251 and task 3 -DETAIL: query string: "SELECT "pg_merge_job_1250.task_000019".intermediate_column_1250_0, "pg_merge_job_1250.task_000019".intermediate_column_1250_1, "pg_merge_job_1250.task_000019".intermediate_column_1250_2, "pg_merge_job_1250.task_000019".intermediate_column_1250_3, "pg_merge_job_1250.task_000019".intermediate_column_1250_4 FROM (pg_merge_job_1250.task_000019 "pg_merge_job_1250.task_000019" JOIN part_102019 part ON (("pg_merge_job_1250.task_000019".intermediate_column_1250_0 = part.p_partkey))) WHERE (part.p_size > 8)" +DETAIL: query string: "SELECT "pg_merge_job_1250.task_000019".intermediate_column_1250_0, "pg_merge_job_1250.task_000019".intermediate_column_1250_1, "pg_merge_job_1250.task_000019".intermediate_column_1250_2, "pg_merge_job_1250.task_000019".intermediate_column_1250_3, "pg_merge_job_1250.task_000019".intermediate_column_1250_4 FROM (pg_merge_job_1250.task_000019 "pg_merge_job_1250.task_000019" JOIN part_290010 part ON (("pg_merge_job_1250.task_000019".intermediate_column_1250_0 = part.p_partkey))) WHERE (part.p_size > 8)" DEBUG: generated sql query for job 1251 and task 6 -DETAIL: query string: "SELECT "pg_merge_job_1250.task_000026".intermediate_column_1250_0, "pg_merge_job_1250.task_000026".intermediate_column_1250_1, "pg_merge_job_1250.task_000026".intermediate_column_1250_2, "pg_merge_job_1250.task_000026".intermediate_column_1250_3, "pg_merge_job_1250.task_000026".intermediate_column_1250_4 FROM (pg_merge_job_1250.task_000026 "pg_merge_job_1250.task_000026" JOIN part_102044 part ON (("pg_merge_job_1250.task_000026".intermediate_column_1250_0 = part.p_partkey))) WHERE (part.p_size > 8)" +DETAIL: query string: "SELECT "pg_merge_job_1250.task_000026".intermediate_column_1250_0, "pg_merge_job_1250.task_000026".intermediate_column_1250_1, "pg_merge_job_1250.task_000026".intermediate_column_1250_2, "pg_merge_job_1250.task_000026".intermediate_column_1250_3, "pg_merge_job_1250.task_000026".intermediate_column_1250_4 FROM (pg_merge_job_1250.task_000026 "pg_merge_job_1250.task_000026" JOIN part_280002 part ON (("pg_merge_job_1250.task_000026".intermediate_column_1250_0 = part.p_partkey))) WHERE (part.p_size > 8)" DEBUG: pruning merge fetch taskId 1 DETAIL: Creating dependency on merge taskId 19 DEBUG: pruning merge fetch taskId 4 @@ -88,11 +90,11 @@ DEBUG: join prunable for intervals [1001,2000] and [6001,7000] DEBUG: join prunable for intervals [6001,7000] and [1,1000] DEBUG: join prunable for intervals [6001,7000] and [1001,2000] DEBUG: generated sql query for job 1252 and task 3 -DETAIL: query string: "SELECT "pg_merge_job_1251.task_000007".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000007 "pg_merge_job_1251.task_000007" JOIN customer_102017 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000007".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000007".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000007".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1, "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 LIMIT 30::bigint" +DETAIL: query string: "SELECT "pg_merge_job_1251.task_000007".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000007 "pg_merge_job_1251.task_000007" JOIN customer_290008 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000007".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000007".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000007".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1, "pg_merge_job_1251.task_000007".intermediate_column_1251_0, "pg_merge_job_1251.task_000007".intermediate_column_1251_1 LIMIT 30::bigint" DEBUG: generated sql query for job 1252 and task 6 -DETAIL: query string: "SELECT "pg_merge_job_1251.task_000010".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000010 "pg_merge_job_1251.task_000010" JOIN customer_102043 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000010".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000010".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000010".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1, "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 LIMIT 30::bigint" +DETAIL: query string: "SELECT "pg_merge_job_1251.task_000010".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000010 "pg_merge_job_1251.task_000010" JOIN customer_280001 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000010".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000010".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000010".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1, "pg_merge_job_1251.task_000010".intermediate_column_1251_0, "pg_merge_job_1251.task_000010".intermediate_column_1251_1 LIMIT 30::bigint" DEBUG: generated sql query for job 1252 and task 9 -DETAIL: query string: "SELECT "pg_merge_job_1251.task_000013".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000013 "pg_merge_job_1251.task_000013" JOIN customer_102042 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000013".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000013".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000013".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1, "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 LIMIT 30::bigint" +DETAIL: query string: "SELECT "pg_merge_job_1251.task_000013".intermediate_column_1251_0 AS l_partkey, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 AS o_orderkey, count(*) AS count FROM (pg_merge_job_1251.task_000013 "pg_merge_job_1251.task_000013" JOIN customer_280000 customer ON ((customer.c_custkey = "pg_merge_job_1251.task_000013".intermediate_column_1251_4))) WHERE ((("pg_merge_job_1251.task_000013".intermediate_column_1251_2 > 5.0) OR ("pg_merge_job_1251.task_000013".intermediate_column_1251_3 > 1200.0)) AND (customer.c_acctbal < 5000.0)) GROUP BY "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 ORDER BY "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1, "pg_merge_job_1251.task_000013".intermediate_column_1251_0, "pg_merge_job_1251.task_000013".intermediate_column_1251_1 LIMIT 30::bigint" DEBUG: pruning merge fetch taskId 1 DETAIL: Creating dependency on merge taskId 7 DEBUG: pruning merge fetch taskId 4 @@ -100,8 +102,8 @@ DETAIL: Creating dependency on merge taskId 10 DEBUG: pruning merge fetch taskId 7 DETAIL: Creating dependency on merge taskId 13 DEBUG: assigned task 9 to node localhost:57637 -DEBUG: assigned task 3 to node localhost:57638 -DEBUG: assigned task 6 to node localhost:57637 +DEBUG: assigned task 6 to node localhost:57638 +DEBUG: assigned task 3 to node localhost:57637 DEBUG: completed cleanup query for job 1252 on node "localhost:57638" DEBUG: completed cleanup query for job 1252 on node "localhost:57637" DEBUG: completed cleanup query for job 1251 on node "localhost:57638" @@ -156,29 +158,29 @@ ORDER BY l_partkey, o_orderkey; DEBUG: StartTransactionCommand DEBUG: generated sql query for job 1253 and task 2 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102009 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290000 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 4 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102010 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290001 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 6 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102011 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290002 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 8 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102012 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290003 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 10 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102013 lineitem WHERE (l_quantity < 5.0)" +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290004 lineitem WHERE (l_quantity < 5.0)" DEBUG: generated sql query for job 1253 and task 12 -DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_102014 lineitem WHERE (l_quantity < 5.0)" -DEBUG: assigned task 4 to node localhost:57637 -DEBUG: assigned task 2 to node localhost:57638 -DEBUG: assigned task 8 to node localhost:57637 -DEBUG: assigned task 6 to node localhost:57638 -DEBUG: assigned task 12 to node localhost:57637 -DEBUG: assigned task 10 to node localhost:57638 +DETAIL: query string: "SELECT l_partkey, l_suppkey FROM lineitem_290005 lineitem WHERE (l_quantity < 5.0)" +DEBUG: assigned task 2 to node localhost:57637 +DEBUG: assigned task 4 to node localhost:57638 +DEBUG: assigned task 6 to node localhost:57637 +DEBUG: assigned task 8 to node localhost:57638 +DEBUG: assigned task 10 to node localhost:57637 +DEBUG: assigned task 12 to node localhost:57638 DEBUG: generated sql query for job 1254 and task 2 -DETAIL: query string: "SELECT o_orderkey, o_shippriority FROM orders_102015 orders WHERE (o_totalprice <> 4.0)" +DETAIL: query string: "SELECT o_orderkey, o_shippriority FROM orders_290006 orders WHERE (o_totalprice <> 4.0)" DEBUG: generated sql query for job 1254 and task 4 -DETAIL: query string: "SELECT o_orderkey, o_shippriority FROM orders_102016 orders WHERE (o_totalprice <> 4.0)" -DEBUG: assigned task 4 to node localhost:57637 -DEBUG: assigned task 2 to node localhost:57638 +DETAIL: query string: "SELECT o_orderkey, o_shippriority FROM orders_290007 orders WHERE (o_totalprice <> 4.0)" +DEBUG: assigned task 2 to node localhost:57637 +DEBUG: assigned task 4 to node localhost:57638 DEBUG: join prunable for task partitionId 0 and 1 DEBUG: join prunable for task partitionId 0 and 2 DEBUG: join prunable for task partitionId 0 and 3 diff --git a/src/test/regress/expected/multi_large_table_pruning.out b/src/test/regress/expected/multi_large_table_pruning.out index d20ae46bc..cb377cf5f 100644 --- a/src/test/regress/expected/multi_large_table_pruning.out +++ b/src/test/regress/expected/multi_large_table_pruning.out @@ -4,6 +4,8 @@ -- Tests covering partition and join-pruning for large table joins. Note that we -- set executor type to task tracker executor here, as we cannot run repartition -- jobs with real time executor. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 700000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 700000; SET citus.large_table_shard_count TO 2; SET client_min_messages TO DEBUG2; SET citus.task_executor_type TO 'task-tracker'; @@ -40,8 +42,8 @@ FROM WHERE o_custkey = c_custkey AND o_orderkey < 0; -DEBUG: predicate pruning for shardId 102015 -DEBUG: predicate pruning for shardId 102016 +DEBUG: predicate pruning for shardId 290006 +DEBUG: predicate pruning for shardId 290007 count ------- @@ -56,9 +58,9 @@ FROM WHERE o_custkey = c_custkey AND c_custkey < 0; -DEBUG: predicate pruning for shardId 102017 -DEBUG: predicate pruning for shardId 102043 -DEBUG: predicate pruning for shardId 102042 +DEBUG: predicate pruning for shardId 290008 +DEBUG: predicate pruning for shardId 280001 +DEBUG: predicate pruning for shardId 280000 count ------- @@ -115,12 +117,12 @@ FROM WHERE l_partkey = c_nationkey AND l_orderkey < 0; -DEBUG: predicate pruning for shardId 102009 -DEBUG: predicate pruning for shardId 102010 -DEBUG: predicate pruning for shardId 102011 -DEBUG: predicate pruning for shardId 102012 -DEBUG: predicate pruning for shardId 102013 -DEBUG: predicate pruning for shardId 102014 +DEBUG: predicate pruning for shardId 290000 +DEBUG: predicate pruning for shardId 290001 +DEBUG: predicate pruning for shardId 290002 +DEBUG: predicate pruning for shardId 290003 +DEBUG: predicate pruning for shardId 290004 +DEBUG: predicate pruning for shardId 290005 count ------- diff --git a/src/test/regress/expected/multi_large_table_task_assignment.out b/src/test/regress/expected/multi_large_table_task_assignment.out index 74747311a..c5a95c149 100644 --- a/src/test/regress/expected/multi_large_table_task_assignment.out +++ b/src/test/regress/expected/multi_large_table_task_assignment.out @@ -5,6 +5,8 @@ -- and dual hash repartition joins. The tests also cover task assignment propagation -- from a sql task to its depended tasks. Note that we set the executor type to task -- tracker executor here, as we cannot run repartition jobs with real time executor. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 710000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 710000; BEGIN; SET client_min_messages TO DEBUG3; DEBUG: CommitTransactionCommand @@ -25,8 +27,8 @@ FROM WHERE o_custkey = c_custkey; DEBUG: StartTransactionCommand -DEBUG: assigned task 4 to node localhost:57637 -DEBUG: assigned task 2 to node localhost:57638 +DEBUG: assigned task 2 to node localhost:57637 +DEBUG: assigned task 4 to node localhost:57638 DEBUG: join prunable for intervals [1,1000] and [1001,2000] DEBUG: join prunable for intervals [1,1000] and [6001,7000] DEBUG: join prunable for intervals [1001,2000] and [1,1000] @@ -40,8 +42,8 @@ DETAIL: Creating dependency on merge taskId 8 DEBUG: pruning merge fetch taskId 7 DETAIL: Creating dependency on merge taskId 11 DEBUG: assigned task 9 to node localhost:57637 -DEBUG: assigned task 3 to node localhost:57638 -DEBUG: assigned task 6 to node localhost:57637 +DEBUG: assigned task 6 to node localhost:57638 +DEBUG: assigned task 3 to node localhost:57637 DEBUG: CommitTransactionCommand count ------- @@ -65,11 +67,11 @@ WHERE o_orderkey = l_orderkey; DEBUG: StartTransactionCommand DEBUG: assigned task 15 to node localhost:57637 -DEBUG: assigned task 3 to node localhost:57638 +DEBUG: assigned task 9 to node localhost:57638 DEBUG: assigned task 18 to node localhost:57637 -DEBUG: assigned task 6 to node localhost:57638 -DEBUG: assigned task 9 to node localhost:57637 DEBUG: assigned task 12 to node localhost:57638 +DEBUG: assigned task 3 to node localhost:57637 +DEBUG: assigned task 6 to node localhost:57638 DEBUG: join prunable for intervals [1,2496] and [2497,4964] DEBUG: join prunable for intervals [1,2496] and [4965,5986] DEBUG: join prunable for intervals [1,2496] and [8997,11554] @@ -114,10 +116,10 @@ DEBUG: pruning merge fetch taskId 19 DETAIL: Creating dependency on merge taskId 47 DEBUG: pruning merge fetch taskId 22 DETAIL: Creating dependency on merge taskId 54 -DEBUG: assigned task 6 to node localhost:57637 -DEBUG: assigned task 3 to node localhost:57638 -DEBUG: assigned task 12 to node localhost:57637 -DEBUG: assigned task 9 to node localhost:57638 +DEBUG: assigned task 3 to node localhost:57637 +DEBUG: assigned task 6 to node localhost:57638 +DEBUG: assigned task 9 to node localhost:57637 +DEBUG: assigned task 12 to node localhost:57638 DEBUG: assigned task 18 to node localhost:57637 DEBUG: assigned task 24 to node localhost:57638 DEBUG: propagating assignment from merge task 40 to constrained sql task 15 @@ -154,15 +156,15 @@ FROM WHERE l_partkey = c_nationkey; DEBUG: StartTransactionCommand -DEBUG: assigned task 4 to node localhost:57637 -DEBUG: assigned task 2 to node localhost:57638 -DEBUG: assigned task 8 to node localhost:57637 -DEBUG: assigned task 6 to node localhost:57638 -DEBUG: assigned task 12 to node localhost:57637 -DEBUG: assigned task 10 to node localhost:57638 +DEBUG: assigned task 2 to node localhost:57637 +DEBUG: assigned task 4 to node localhost:57638 DEBUG: assigned task 6 to node localhost:57637 -DEBUG: assigned task 2 to node localhost:57638 -DEBUG: assigned task 4 to node localhost:57637 +DEBUG: assigned task 8 to node localhost:57638 +DEBUG: assigned task 10 to node localhost:57637 +DEBUG: assigned task 12 to node localhost:57638 +DEBUG: assigned task 6 to node localhost:57637 +DEBUG: assigned task 4 to node localhost:57638 +DEBUG: assigned task 2 to node localhost:57637 DEBUG: join prunable for task partitionId 0 and 1 DEBUG: join prunable for task partitionId 0 and 2 DEBUG: join prunable for task partitionId 0 and 3 diff --git a/src/test/regress/expected/multi_limit_clause.out b/src/test/regress/expected/multi_limit_clause.out index 18e1adf20..4ead2585f 100644 --- a/src/test/regress/expected/multi_limit_clause.out +++ b/src/test/regress/expected/multi_limit_clause.out @@ -1,6 +1,8 @@ -- -- MULTI_LIMIT_CLAUSE -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 730000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 730000; -- Display debug messages on limit clause push down. SET client_min_messages TO DEBUG1; -- Check that we can correctly handle the Limit clause in distributed queries. diff --git a/src/test/regress/expected/multi_limit_clause_approximate.out b/src/test/regress/expected/multi_limit_clause_approximate.out index d1bad570e..7008a52a8 100644 --- a/src/test/regress/expected/multi_limit_clause_approximate.out +++ b/src/test/regress/expected/multi_limit_clause_approximate.out @@ -1,6 +1,8 @@ -- -- MULTI_LIMIT_CLAUSE_APPROXIMATE -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 720000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 720000; -- Display debug messages on limit clause push down. SET client_min_messages TO DEBUG1; -- We first look at results with limit optimization disabled. This first query diff --git a/src/test/regress/expected/multi_master_protocol.out b/src/test/regress/expected/multi_master_protocol.out index 3d6e627a5..b000b15af 100644 --- a/src/test/regress/expected/multi_master_protocol.out +++ b/src/test/regress/expected/multi_master_protocol.out @@ -2,6 +2,8 @@ -- MULTI_MASTER_PROTOCOL -- -- Tests that check the metadata returned by the master node. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 740000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 740000; SELECT part_storage_type, part_key, part_replica_count, part_max_size, part_placement_policy FROM master_get_table_metadata('lineitem'); part_storage_type | part_key | part_replica_count | part_max_size | part_placement_policy @@ -20,7 +22,7 @@ SELECT * FROM master_get_table_ddl_events('lineitem'); SELECT * FROM master_get_new_shardid(); master_get_new_shardid ------------------------ - 102008 + 740000 (1 row) SELECT * FROM master_get_local_first_candidate_nodes(); diff --git a/src/test/regress/expected/multi_modifications.out b/src/test/regress/expected/multi_modifications.out index 84b519435..4a707f733 100644 --- a/src/test/regress/expected/multi_modifications.out +++ b/src/test/regress/expected/multi_modifications.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 750000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 750000; -- =================================================================== -- test end-to-end modification functionality -- =================================================================== @@ -93,7 +95,7 @@ INSERT INTO append_partitioned VALUES (414123, 'AAPL', 9580, '2004-10-19 10:23:5 SET client_min_messages TO 'DEBUG2'; SET citus.task_executor_type TO 'real-time'; SELECT * FROM range_partitioned WHERE id = 32743; -DEBUG: predicate pruning for shardId 103084 +DEBUG: predicate pruning for shardId 750004 DEBUG: Plan is router executable id | symbol | bidder_id | placed_at | kind | limit_price -------+--------+-----------+--------------------------+------+------------- @@ -101,7 +103,7 @@ DEBUG: Plan is router executable (1 row) SELECT * FROM append_partitioned WHERE id = 414123; -DEBUG: predicate pruning for shardId 103086 +DEBUG: predicate pruning for shardId 750006 DEBUG: Plan is router executable id | symbol | bidder_id | placed_at | kind | limit_price --------+--------+-----------+--------------------------+------+------------- @@ -275,7 +277,7 @@ WHERE nodename = 'localhost' AND -- Fourth: Perform the same INSERT (primary key violation) INSERT INTO limit_orders VALUES (275, 'ADR', 140, '2007-07-02 16:32:15', 'sell', 43.67); WARNING: Bad result from localhost:57638 -DETAIL: Remote message: duplicate key value violates unique constraint "limit_orders_pkey_103081" +DETAIL: Remote message: duplicate key value violates unique constraint "limit_orders_pkey_750001" -- Last: Verify the insert worked but the placement with the PK violation is now unhealthy SELECT count(*) FROM limit_orders WHERE id = 275; count diff --git a/src/test/regress/expected/multi_null_minmax_value_pruning.out b/src/test/regress/expected/multi_null_minmax_value_pruning.out index 80795483c..fc3321652 100644 --- a/src/test/regress/expected/multi_null_minmax_value_pruning.out +++ b/src/test/regress/expected/multi_null_minmax_value_pruning.out @@ -3,16 +3,18 @@ -- -- This test checks that we can handle null min/max values in shard statistics -- and that we don't partition or join prune shards that have null values. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 760000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 760000; SET client_min_messages TO DEBUG2; -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; -SELECT shardminvalue, shardmaxvalue from pg_dist_shard WHERE shardid = 102009; +SELECT shardminvalue, shardmaxvalue from pg_dist_shard WHERE shardid = 290000; shardminvalue | shardmaxvalue ---------------+--------------- 1 | 2496 (1 row) -SELECT shardminvalue, shardmaxvalue from pg_dist_shard WHERE shardid = 102010; +SELECT shardminvalue, shardmaxvalue from pg_dist_shard WHERE shardid = 290001; shardminvalue | shardmaxvalue ---------------+--------------- 2497 | 4964 @@ -21,24 +23,24 @@ SELECT shardminvalue, shardmaxvalue from pg_dist_shard WHERE shardid = 102010; -- Check that partition and join pruning works when min/max values exist -- Adding l_orderkey = 1 to make the query not router executable SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 9030 or l_orderkey = 1; -DEBUG: predicate pruning for shardId 102010 -DEBUG: predicate pruning for shardId 102011 -DEBUG: predicate pruning for shardId 102013 -DEBUG: predicate pruning for shardId 102014 +DEBUG: predicate pruning for shardId 290001 +DEBUG: predicate pruning for shardId 290002 +DEBUG: predicate pruning for shardId 290004 +DEBUG: predicate pruning for shardId 290005 l_orderkey | l_linenumber | l_shipdate ------------+--------------+------------ - 9030 | 1 | 09-02-1998 - 9030 | 2 | 08-19-1998 - 9030 | 3 | 08-27-1998 - 9030 | 4 | 07-20-1998 - 9030 | 5 | 09-29-1998 - 9030 | 6 | 09-03-1998 1 | 1 | 03-13-1996 1 | 2 | 04-12-1996 1 | 3 | 01-29-1996 1 | 4 | 04-21-1996 1 | 5 | 03-30-1996 1 | 6 | 01-30-1996 + 9030 | 1 | 09-02-1998 + 9030 | 2 | 08-19-1998 + 9030 | 3 | 08-27-1998 + 9030 | 4 | 07-20-1998 + 9030 | 5 | 09-29-1998 + 9030 | 6 | 09-03-1998 (12 rows) SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem, orders @@ -56,12 +58,12 @@ DEBUG: join prunable for intervals [13921,14947] and [1,5986] -- Now set the minimum value for a shard to null. Then check that we don't apply -- partition or join pruning for the shard with null min value. -UPDATE pg_dist_shard SET shardminvalue = NULL WHERE shardid = 102009; +UPDATE pg_dist_shard SET shardminvalue = NULL WHERE shardid = 290000; SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 9030; -DEBUG: predicate pruning for shardId 102010 -DEBUG: predicate pruning for shardId 102011 -DEBUG: predicate pruning for shardId 102013 -DEBUG: predicate pruning for shardId 102014 +DEBUG: predicate pruning for shardId 290001 +DEBUG: predicate pruning for shardId 290002 +DEBUG: predicate pruning for shardId 290004 +DEBUG: predicate pruning for shardId 290005 l_orderkey | l_linenumber | l_shipdate ------------+--------------+------------ 9030 | 1 | 09-02-1998 @@ -86,11 +88,11 @@ DEBUG: join prunable for intervals [13921,14947] and [1,5986] -- Next, set the maximum value for another shard to null. Then check that we -- don't apply partition or join pruning for this other shard either. -UPDATE pg_dist_shard SET shardmaxvalue = NULL WHERE shardid = 102010; +UPDATE pg_dist_shard SET shardmaxvalue = NULL WHERE shardid = 290001; SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 9030; -DEBUG: predicate pruning for shardId 102011 -DEBUG: predicate pruning for shardId 102013 -DEBUG: predicate pruning for shardId 102014 +DEBUG: predicate pruning for shardId 290002 +DEBUG: predicate pruning for shardId 290004 +DEBUG: predicate pruning for shardId 290005 l_orderkey | l_linenumber | l_shipdate ------------+--------------+------------ 9030 | 1 | 09-02-1998 @@ -114,12 +116,12 @@ DEBUG: join prunable for intervals [13921,14947] and [1,5986] -- Last, set the minimum value to 0 and check that we don't treat it as null. We -- should apply partition and join pruning for this shard now. -UPDATE pg_dist_shard SET shardminvalue = '0' WHERE shardid = 102009; +UPDATE pg_dist_shard SET shardminvalue = '0' WHERE shardid = 290000; SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 9030; -DEBUG: predicate pruning for shardId 102009 -DEBUG: predicate pruning for shardId 102011 -DEBUG: predicate pruning for shardId 102013 -DEBUG: predicate pruning for shardId 102014 +DEBUG: predicate pruning for shardId 290000 +DEBUG: predicate pruning for shardId 290002 +DEBUG: predicate pruning for shardId 290004 +DEBUG: predicate pruning for shardId 290005 l_orderkey | l_linenumber | l_shipdate ------------+--------------+------------ 9030 | 1 | 09-02-1998 @@ -143,6 +145,6 @@ DEBUG: join prunable for intervals [13921,14947] and [1,5986] (1 row) -- Set minimum and maximum values for two shards back to their original values -UPDATE pg_dist_shard SET shardminvalue = '1' WHERE shardid = 102009; -UPDATE pg_dist_shard SET shardmaxvalue = '4964' WHERE shardid = 102010; +UPDATE pg_dist_shard SET shardminvalue = '1' WHERE shardid = 290000; +UPDATE pg_dist_shard SET shardmaxvalue = '4964' WHERE shardid = 290001; SET client_min_messages TO NOTICE; diff --git a/src/test/regress/expected/multi_partition_pruning.out b/src/test/regress/expected/multi_partition_pruning.out index 0a3cade99..3d305b0e7 100644 --- a/src/test/regress/expected/multi_partition_pruning.out +++ b/src/test/regress/expected/multi_partition_pruning.out @@ -3,28 +3,30 @@ -- -- Tests to verify that we correctly prune unreferenced shards. For this, we -- need to increase the logging verbosity of messages displayed on the client. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 770000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 770000; SET citus.explain_distributed_queries TO off; SET client_min_messages TO DEBUG2; -- Adding additional l_orderkey = 1 to make this query not router executable SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 9030 or l_orderkey = 1; -DEBUG: predicate pruning for shardId 102010 -DEBUG: predicate pruning for shardId 102011 -DEBUG: predicate pruning for shardId 102013 -DEBUG: predicate pruning for shardId 102014 +DEBUG: predicate pruning for shardId 290001 +DEBUG: predicate pruning for shardId 290002 +DEBUG: predicate pruning for shardId 290004 +DEBUG: predicate pruning for shardId 290005 l_orderkey | l_linenumber | l_shipdate ------------+--------------+------------ - 9030 | 1 | 09-02-1998 - 9030 | 2 | 08-19-1998 - 9030 | 3 | 08-27-1998 - 9030 | 4 | 07-20-1998 - 9030 | 5 | 09-29-1998 - 9030 | 6 | 09-03-1998 1 | 1 | 03-13-1996 1 | 2 | 04-12-1996 1 | 3 | 01-29-1996 1 | 4 | 04-21-1996 1 | 5 | 03-30-1996 1 | 6 | 01-30-1996 + 9030 | 1 | 09-02-1998 + 9030 | 2 | 08-19-1998 + 9030 | 3 | 08-27-1998 + 9030 | 4 | 07-20-1998 + 9030 | 5 | 09-29-1998 + 9030 | 6 | 09-03-1998 (12 rows) -- We use the l_linenumber field for the following aggregations. We need to use @@ -33,9 +35,9 @@ DEBUG: predicate pruning for shardId 102014 -- trigger the the creation of toasted tables and indexes. This in turn prints -- non-deterministic debug messages. To avoid this chain, we use l_linenumber. SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030; -DEBUG: predicate pruning for shardId 102009 -DEBUG: predicate pruning for shardId 102010 -DEBUG: predicate pruning for shardId 102011 +DEBUG: predicate pruning for shardId 290000 +DEBUG: predicate pruning for shardId 290001 +DEBUG: predicate pruning for shardId 290002 sum | avg -------+-------------------- 17999 | 3.0189533713518953 @@ -43,7 +45,7 @@ DEBUG: predicate pruning for shardId 102011 SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem WHERE (l_orderkey < 4000 OR l_orderkey > 9030); -DEBUG: predicate pruning for shardId 102011 +DEBUG: predicate pruning for shardId 290002 sum | avg -------+-------------------- 30184 | 3.0159872102318145 @@ -51,12 +53,12 @@ DEBUG: predicate pruning for shardId 102011 -- The following query should prune out all shards and return empty results SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem WHERE l_orderkey > 20000; -DEBUG: predicate pruning for shardId 102009 -DEBUG: predicate pruning for shardId 102010 -DEBUG: predicate pruning for shardId 102011 -DEBUG: predicate pruning for shardId 102012 -DEBUG: predicate pruning for shardId 102013 -DEBUG: predicate pruning for shardId 102014 +DEBUG: predicate pruning for shardId 290000 +DEBUG: predicate pruning for shardId 290001 +DEBUG: predicate pruning for shardId 290002 +DEBUG: predicate pruning for shardId 290003 +DEBUG: predicate pruning for shardId 290004 +DEBUG: predicate pruning for shardId 290005 sum | avg -----+----- | diff --git a/src/test/regress/expected/multi_prepare_plsql.out b/src/test/regress/expected/multi_prepare_plsql.out index e76482076..1932ce862 100644 --- a/src/test/regress/expected/multi_prepare_plsql.out +++ b/src/test/regress/expected/multi_prepare_plsql.out @@ -4,6 +4,8 @@ -- Many of the queries are taken from other regression test files -- and converted into both plain SQL and PL/pgsql functions, which -- use prepared statements internally. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 780000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 780000; CREATE FUNCTION sql_test_no_1() RETURNS bigint AS ' SELECT count(*) diff --git a/src/test/regress/expected/multi_prepare_sql.out b/src/test/regress/expected/multi_prepare_sql.out index b0d097a3f..3a54596a6 100644 --- a/src/test/regress/expected/multi_prepare_sql.out +++ b/src/test/regress/expected/multi_prepare_sql.out @@ -4,6 +4,8 @@ -- Tests covering PREPARE statements. Many of the queries are -- taken from other regression test files and converted into -- prepared statements. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 790000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 790000; PREPARE prepared_test_1 AS SELECT count(*) diff --git a/src/test/regress/expected/multi_prune_shard_list.out b/src/test/regress/expected/multi_prune_shard_list.out index 934639051..ea8594c24 100644 --- a/src/test/regress/expected/multi_prune_shard_list.out +++ b/src/test/regress/expected/multi_prune_shard_list.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 800000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 800000; -- =================================================================== -- create test functions -- =================================================================== @@ -47,28 +49,28 @@ SELECT master_create_worker_shards('pruning', 4, 1); SELECT prune_using_no_values('pruning'); prune_using_no_values ------------------------------- - {103070,103071,103072,103073} + {800000,800001,800002,800003} (1 row) -- with a single value, expect a single shard SELECT prune_using_single_value('pruning', 'tomato'); prune_using_single_value -------------------------- - {103072} + {800002} (1 row) -- the above is true even if that value is null SELECT prune_using_single_value('pruning', NULL); prune_using_single_value -------------------------- - {103072} + {800002} (1 row) -- build an OR clause and expect more than one sahrd SELECT prune_using_either_value('pruning', 'tomato', 'petunia'); prune_using_either_value -------------------------- - {103071,103072} + {800001,800002} (1 row) -- an AND clause with incompatible values returns no shards @@ -82,7 +84,7 @@ SELECT prune_using_both_values('pruning', 'tomato', 'petunia'); SELECT prune_using_both_values('pruning', 'tomato', 'rose'); prune_using_both_values ------------------------- - {103072} + {800002} (1 row) -- unit test of the equality expression generation code @@ -96,7 +98,7 @@ SELECT debug_equality_expression('pruning'); SELECT print_sorted_shard_intervals('pruning'); print_sorted_shard_intervals ------------------------------- - {103070,103071,103072,103073} + {800000,800001,800002,800003} (1 row) -- update only min value for one shard @@ -104,7 +106,7 @@ UPDATE pg_dist_shard set shardminvalue = NULL, shardmaxvalue = NULL WHERE shardi SELECT print_sorted_shard_intervals('pruning'); print_sorted_shard_intervals ------------------------------- - {103070,103072,103073,103071} + {800000,800001,800002,800003} (1 row) -- now lets have one more shard without min/max values @@ -112,7 +114,7 @@ UPDATE pg_dist_shard set shardminvalue = NULL, shardmaxvalue = NULL WHERE shardi SELECT print_sorted_shard_intervals('pruning'); print_sorted_shard_intervals ------------------------------- - {103070,103073,103071,103072} + {800000,800001,800002,800003} (1 row) -- now lets have one more shard without min/max values @@ -120,7 +122,7 @@ UPDATE pg_dist_shard set shardminvalue = NULL, shardmaxvalue = NULL WHERE shardi SELECT print_sorted_shard_intervals('pruning'); print_sorted_shard_intervals ------------------------------- - {103073,103070,103071,103072} + {800000,800001,800002,800003} (1 row) -- all shard placements are uninitialized @@ -128,7 +130,7 @@ UPDATE pg_dist_shard set shardminvalue = NULL, shardmaxvalue = NULL WHERE shardi SELECT print_sorted_shard_intervals('pruning'); print_sorted_shard_intervals ------------------------------- - {103070,103071,103072,103073} + {800000,800001,800002,800003} (1 row) -- create range distributed table observe shard pruning @@ -143,25 +145,25 @@ SELECT master_create_distributed_table('pruning_range', 'species', 'range'); SELECT master_create_empty_shard('pruning_range'); master_create_empty_shard --------------------------- - 103074 + 800004 (1 row) SELECT master_create_empty_shard('pruning_range'); master_create_empty_shard --------------------------- - 103075 + 800005 (1 row) SELECT master_create_empty_shard('pruning_range'); master_create_empty_shard --------------------------- - 103076 + 800006 (1 row) SELECT master_create_empty_shard('pruning_range'); master_create_empty_shard --------------------------- - 103077 + 800007 (1 row) -- now the comparison is done via the partition column type, which is text @@ -173,7 +175,7 @@ UPDATE pg_dist_shard SET shardminvalue = 'g', shardmaxvalue = 'h' WHERE shardid SELECT print_sorted_shard_intervals('pruning_range'); print_sorted_shard_intervals ------------------------------- - {103074,103075,103076,103077} + {800004,800005,800006,800007} (1 row) -- update only min value for one shard @@ -181,7 +183,7 @@ UPDATE pg_dist_shard set shardminvalue = NULL, shardmaxvalue = NULL WHERE shardi SELECT print_sorted_shard_intervals('pruning_range'); print_sorted_shard_intervals ------------------------------- - {103074,103076,103077,103075} + {800004,800005,800006,800007} (1 row) -- now lets have one more shard without min/max values @@ -189,7 +191,7 @@ UPDATE pg_dist_shard set shardminvalue = NULL, shardmaxvalue = NULL WHERE shardi SELECT print_sorted_shard_intervals('pruning_range'); print_sorted_shard_intervals ------------------------------- - {103074,103077,103075,103076} + {800004,800005,800006,800007} (1 row) -- now lets have one more shard without min/max values @@ -197,7 +199,7 @@ UPDATE pg_dist_shard set shardminvalue = NULL, shardmaxvalue = NULL WHERE shardi SELECT print_sorted_shard_intervals('pruning_range'); print_sorted_shard_intervals ------------------------------- - {103077,103074,103075,103076} + {800004,800005,800006,800007} (1 row) -- all shard placements are uninitialized @@ -205,6 +207,6 @@ UPDATE pg_dist_shard set shardminvalue = NULL, shardmaxvalue = NULL WHERE shardi SELECT print_sorted_shard_intervals('pruning_range'); print_sorted_shard_intervals ------------------------------- - {103074,103075,103076,103077} + {800004,800005,800006,800007} (1 row) diff --git a/src/test/regress/expected/multi_query_directory_cleanup.out b/src/test/regress/expected/multi_query_directory_cleanup.out index a72a3a06d..65caed2e9 100644 --- a/src/test/regress/expected/multi_query_directory_cleanup.out +++ b/src/test/regress/expected/multi_query_directory_cleanup.out @@ -5,6 +5,8 @@ -- on the master node for final processing. When the query completes or fails, -- the resource owner should automatically clean up these intermediate query -- result files. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 810000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 810000; BEGIN; -- pg_ls_dir() displays jobids. We explicitly set the jobId sequence -- here so that the regression output becomes independent of the diff --git a/src/test/regress/expected/multi_repair_shards.out b/src/test/regress/expected/multi_repair_shards.out index bf76df9c5..29a5f96ec 100644 --- a/src/test/regress/expected/multi_repair_shards.out +++ b/src/test/regress/expected/multi_repair_shards.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 820000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 820000; -- =================================================================== -- test shard repair functionality -- =================================================================== diff --git a/src/test/regress/expected/multi_repartitioned_subquery_udf.out b/src/test/regress/expected/multi_repartitioned_subquery_udf.out index 288ee5427..39484ca4e 100644 --- a/src/test/regress/expected/multi_repartitioned_subquery_udf.out +++ b/src/test/regress/expected/multi_repartitioned_subquery_udf.out @@ -1,6 +1,8 @@ -- -- MULTI_REPARTITIONED_SUBQUERY_UDF -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 830000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 830000; -- Create UDF in master and workers \c - - - :master_port DROP FUNCTION IF EXISTS median(double precision[]); diff --git a/src/test/regress/expected/multi_router_planner.out b/src/test/regress/expected/multi_router_planner.out index 5c8578484..1971b7778 100644 --- a/src/test/regress/expected/multi_router_planner.out +++ b/src/test/regress/expected/multi_router_planner.out @@ -1,7 +1,8 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 840000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 840000; -- =================================================================== -- test router planner functionality for single shard select queries -- =================================================================== -ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 103300; CREATE TABLE articles_hash ( id bigint NOT NULL, author_id bigint NOT NULL, @@ -105,8 +106,8 @@ DEBUG: Creating router plan DEBUG: Plan is router executable -- first, test zero-shard SELECT, which should return an empty row SELECT COUNT(*) FROM articles_hash WHERE author_id = 1 AND author_id = 2; -DEBUG: predicate pruning for shardId 103300 -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840000 +DEBUG: predicate pruning for shardId 840001 count ------- @@ -116,7 +117,7 @@ DEBUG: predicate pruning for shardId 103301 -- test simple select for a single row SELECT * FROM articles_hash WHERE author_id = 10 AND id = 50; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+-----------+------------ @@ -126,7 +127,7 @@ DEBUG: Plan is router executable -- get all titles by a single author SELECT title FROM articles_hash WHERE author_id = 10; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable title ------------ @@ -142,7 +143,7 @@ SELECT title, word_count FROM articles_hash WHERE author_id = 10 ORDER BY word_count DESC NULLS LAST; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable title | word_count ------------+------------ @@ -159,7 +160,7 @@ SELECT title, id FROM articles_hash ORDER BY id LIMIT 2; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable title | id ---------+---- @@ -172,7 +173,7 @@ DEBUG: Plan is router executable SELECT title, author_id FROM articles_hash WHERE author_id = 7 OR author_id = 8 ORDER BY author_id ASC, id; -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 title | author_id -------------+----------- aseptic | 7 @@ -190,7 +191,7 @@ DEBUG: predicate pruning for shardId 103301 -- same query is router executable with no order by SELECT title, author_id FROM articles_hash WHERE author_id = 7 OR author_id = 8; -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable title | author_id -------------+----------- @@ -222,7 +223,7 @@ SELECT author_id, sum(word_count) AS corpus_size FROM articles_hash HAVING sum(word_count) > 1000 ORDER BY sum(word_count) DESC; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable author_id | corpus_size -----------+------------- @@ -320,7 +321,7 @@ SELECT * FROM articles_hash WHERE author_id = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -336,7 +337,7 @@ DEBUG: Plan is router executable SELECT * FROM articles_hash WHERE author_id = 1 OR author_id = 17; -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -366,7 +367,7 @@ SELECT id as article_id, word_count * id as random_value FROM articles_hash WHERE author_id = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable article_id | random_value ------------+-------------- @@ -385,7 +386,7 @@ SELECT a.author_id as first_author, b.word_count as second_word_count WHERE a.author_id = 10 and a.author_id = b.author_id LIMIT 3; DEBUG: push down of limit count: 3 -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: join prunable for intervals [-2147483648,-1] and [0,2147483647] DEBUG: Plan is router executable first_author | second_word_count @@ -401,7 +402,7 @@ SELECT a.author_id as first_author, b.word_count as second_word_count WHERE a.author_id = 10 and a.author_id = b.author_id LIMIT 3; DEBUG: push down of limit count: 3 -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 first_author | second_word_count --------------+------------------- 10 | 19519 @@ -415,7 +416,7 @@ SELECT * WHERE author_id = 1 LIMIT 3; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+----------+------------ @@ -431,7 +432,7 @@ SELECT * LIMIT 2 OFFSET 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+----------+------------ @@ -447,7 +448,7 @@ SELECT * LIMIT 2 OFFSET 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -462,7 +463,7 @@ SELECT id WHERE author_id = 1 GROUP BY id; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id ---- @@ -478,7 +479,7 @@ SELECT distinct id FROM articles_hash WHERE author_id = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id ---- @@ -494,7 +495,7 @@ SELECT avg(word_count) FROM articles_hash WHERE author_id = 2; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103300 +DEBUG: predicate pruning for shardId 840000 DEBUG: Plan is router executable avg -------------------- @@ -507,7 +508,7 @@ SELECT max(word_count) as max, min(word_count) as min, FROM articles_hash WHERE author_id = 2; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103300 +DEBUG: predicate pruning for shardId 840000 DEBUG: Plan is router executable max | min | sum | cnt -------+------+-------+----- @@ -520,7 +521,7 @@ SELECT max(word_count) WHERE author_id = 1 GROUP BY author_id; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable max ------- @@ -566,7 +567,7 @@ SET client_min_messages to 'DEBUG2'; SELECT * FROM articles_hash WHERE author_id = 1 and author_id >= 1; -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -596,7 +597,7 @@ SELECT * FROM articles_hash WHERE author_id = 1 and (id = 1 or id = 41); DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+----------+------------ @@ -609,7 +610,7 @@ SELECT * FROM articles_hash WHERE author_id = 1 and (id = random()::int * 0); DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+-------+------------ @@ -647,7 +648,7 @@ SELECT * FROM articles_hash WHERE author_id = abs(-1); DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -689,7 +690,7 @@ SELECT * FROM articles_hash WHERE author_id = 1 and (id = abs(id - 2)); DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+----------+------------ @@ -714,7 +715,7 @@ SELECT * FROM articles_hash WHERE (author_id = 1) = true; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -730,7 +731,7 @@ SELECT * FROM articles_hash WHERE (author_id = 1) and id between 0 and 20; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+----------+------------ @@ -743,7 +744,7 @@ SELECT * FROM articles_hash WHERE (author_id = 1) and (id = 1 or id = 31) and title like '%s'; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -756,7 +757,7 @@ SELECT * FROM articles_hash WHERE (id = 1 or id = 31) and title like '%s' and (author_id = 1); DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -769,7 +770,7 @@ SELECT * FROM articles_hash WHERE (title like '%s' or title like 'a%') and (author_id = 1); DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -785,7 +786,7 @@ SELECT * FROM articles_hash WHERE (title like '%s' or title like 'a%') and (author_id = 1) and (word_count < 3000 or word_count > 8000); DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+----------+------------ @@ -799,7 +800,7 @@ SELECT LAG(title, 1) over (ORDER BY word_count) prev, title, word_count FROM articles_hash WHERE author_id = 5; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable prev | title | word_count ----------+----------+------------ @@ -815,7 +816,7 @@ SELECT LAG(title, 1) over (ORDER BY word_count) prev, title, word_count WHERE author_id = 5 ORDER BY word_count DESC; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable prev | title | word_count ----------+----------+------------ @@ -830,7 +831,7 @@ SELECT id, MIN(id) over (order by word_count) FROM articles_hash WHERE author_id = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | min ----+----- @@ -845,7 +846,7 @@ SELECT id, word_count, AVG(word_count) over (order by word_count) FROM articles_hash WHERE author_id = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | word_count | avg ----+------------+----------------------- @@ -860,7 +861,7 @@ SELECT word_count, rank() OVER (PARTITION BY author_id ORDER BY word_count) FROM articles_hash WHERE author_id = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable word_count | rank ------------+------ @@ -899,7 +900,7 @@ SELECT WHERE author_id = 5; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable c --- @@ -941,7 +942,7 @@ SELECT * WHERE author_id = 1 ORDER BY id; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -961,7 +962,7 @@ DECLARE test_cursor CURSOR FOR WHERE author_id = 1 ORDER BY id; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable FETCH test_cursor; id | author_id | title | word_count @@ -983,7 +984,7 @@ COPY ( WHERE author_id = 1 ORDER BY id) TO STDOUT; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable 1 1 arsenous 9572 11 1 alamo 1347 @@ -998,14 +999,14 @@ CREATE TEMP TABLE temp_articles_hash as WHERE author_id = 1 ORDER BY id; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable -- router plannable queries may include filter for aggragates SELECT count(*), count(*) FILTER (WHERE id < 3) FROM articles_hash WHERE author_id = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable count | count -------+------- @@ -1028,7 +1029,7 @@ PREPARE author_1_articles as WHERE author_id = 1; EXECUTE author_1_articles; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -1046,7 +1047,7 @@ PREPARE author_articles(int) as WHERE author_id = $1; EXECUTE author_articles(1); DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -1073,7 +1074,7 @@ DEBUG: Creating router plan CONTEXT: SQL statement "SELECT MAX(id) FROM articles_hash ah WHERE author_id = 1" PL/pgSQL function author_articles_max_id() line 5 at SQL statement -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 CONTEXT: SQL statement "SELECT MAX(id) FROM articles_hash ah WHERE author_id = 1" PL/pgSQL function author_articles_max_id() line 5 at SQL statement @@ -1103,7 +1104,7 @@ CONTEXT: SQL statement "SELECT ah.id, ah.word_count FROM articles_hash ah WHERE author_id = 1" PL/pgSQL function author_articles_id_word_count() line 4 at RETURN QUERY -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 CONTEXT: SQL statement "SELECT ah.id, ah.word_count FROM articles_hash ah WHERE author_id = 1" @@ -1118,7 +1119,7 @@ SET citus.task_executor_type to 'task-tracker'; SELECT id FROM articles_hash WHERE author_id = 1; -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 id ---- 1 @@ -1136,7 +1137,7 @@ DEBUG: Plan is router executable SELECT id FROM articles_hash WHERE author_id = 1; -DEBUG: predicate pruning for shardId 103301 +DEBUG: predicate pruning for shardId 840001 id ---- 1 diff --git a/src/test/regress/expected/multi_shard_modify.out b/src/test/regress/expected/multi_shard_modify.out index 03a5ec3a8..59a815659 100644 --- a/src/test/regress/expected/multi_shard_modify.out +++ b/src/test/regress/expected/multi_shard_modify.out @@ -2,6 +2,7 @@ -- MULTI_SHARD_MODIFY -- ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 350000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 350000; -- Create a new hash partitioned multi_shard_modify_test table and stage data into it. CREATE TABLE multi_shard_modify_test ( t_key integer not null, diff --git a/src/test/regress/expected/multi_simple_queries.out b/src/test/regress/expected/multi_simple_queries.out index 9a857511a..6684d578b 100644 --- a/src/test/regress/expected/multi_simple_queries.out +++ b/src/test/regress/expected/multi_simple_queries.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 850000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 850000; -- =================================================================== -- test end-to-end query functionality -- =================================================================== @@ -323,7 +325,7 @@ SELECT * FROM articles WHERE author_id = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103108 +DEBUG: predicate pruning for shardId 850001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -338,7 +340,7 @@ DEBUG: Plan is router executable SELECT * FROM articles WHERE author_id = 1 OR author_id = 17; -DEBUG: predicate pruning for shardId 103108 +DEBUG: predicate pruning for shardId 850001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+--------------+------------ @@ -367,7 +369,7 @@ SELECT id as article_id, word_count * id as random_value FROM articles WHERE author_id = 1; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103108 +DEBUG: predicate pruning for shardId 850001 DEBUG: Plan is router executable article_id | random_value ------------+-------------- @@ -385,7 +387,7 @@ SELECT a.author_id as first_author, b.word_count as second_word_count WHERE a.author_id = 10 and a.author_id = b.author_id LIMIT 3; DEBUG: push down of limit count: 3 -DEBUG: predicate pruning for shardId 103108 +DEBUG: predicate pruning for shardId 850001 DEBUG: join prunable for intervals [-2147483648,-1] and [0,2147483647] DEBUG: Plan is router executable first_author | second_word_count @@ -402,7 +404,7 @@ SELECT a.author_id as first_author, b.word_count as second_word_count WHERE a.author_id = 10 and a.author_id = b.author_id LIMIT 3; DEBUG: push down of limit count: 3 -DEBUG: predicate pruning for shardId 103108 +DEBUG: predicate pruning for shardId 850001 first_author | second_word_count --------------+------------------- 10 | 19519 @@ -416,7 +418,7 @@ SELECT * WHERE author_id = 1 LIMIT 2; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103108 +DEBUG: predicate pruning for shardId 850001 DEBUG: Plan is router executable id | author_id | title | word_count ----+-----------+----------+------------ @@ -432,7 +434,7 @@ SELECT id WHERE author_id = 1 GROUP BY id; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103108 +DEBUG: predicate pruning for shardId 850001 DEBUG: Plan is router executable id ---- @@ -452,7 +454,7 @@ SELECT avg(word_count) FROM articles WHERE author_id = 2; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103107 +DEBUG: predicate pruning for shardId 850000 DEBUG: Plan is router executable avg -------------------- @@ -466,7 +468,7 @@ SELECT max(word_count) as max, min(word_count) as min, FROM articles WHERE author_id = 2; DEBUG: Creating router plan -DEBUG: predicate pruning for shardId 103107 +DEBUG: predicate pruning for shardId 850000 DEBUG: Plan is router executable max | min | sum | cnt -------+------+-------+----- @@ -477,7 +479,7 @@ DEBUG: Plan is router executable SELECT * FROM articles a, articles b WHERE a.id = b.id AND a.author_id = 1; -DEBUG: predicate pruning for shardId 103108 +DEBUG: predicate pruning for shardId 850001 DEBUG: join prunable for task partitionId 0 and 1 DEBUG: join prunable for task partitionId 0 and 2 DEBUG: join prunable for task partitionId 0 and 3 diff --git a/src/test/regress/expected/multi_single_relation_subquery.out b/src/test/regress/expected/multi_single_relation_subquery.out index 47c8c9c82..7e3358819 100644 --- a/src/test/regress/expected/multi_single_relation_subquery.out +++ b/src/test/regress/expected/multi_single_relation_subquery.out @@ -2,6 +2,8 @@ -- MULTI_SINGLE_RELATION_SUBQUERY -- -- This test checks that we are able to run selected set of distributed SQL subqueries. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 860000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 860000; SET citus.task_executor_type TO 'task-tracker'; select number_sum, diff --git a/src/test/regress/expected/multi_table_ddl.out b/src/test/regress/expected/multi_table_ddl.out index 20b39dabb..27400cf5e 100644 --- a/src/test/regress/expected/multi_table_ddl.out +++ b/src/test/regress/expected/multi_table_ddl.out @@ -2,6 +2,8 @@ -- MULTI_TABLE_DDL -- -- Tests around changing the schema and dropping of a distributed table +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 870000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 870000; CREATE TABLE testtableddl(somecol int, distributecol text NOT NULL); SELECT master_create_distributed_table('testtableddl', 'distributecol', 'append'); master_create_distributed_table diff --git a/src/test/regress/expected/multi_task_assignment_policy.out b/src/test/regress/expected/multi_task_assignment_policy.out index 9b6cce828..f571238ad 100644 --- a/src/test/regress/expected/multi_task_assignment_policy.out +++ b/src/test/regress/expected/multi_task_assignment_policy.out @@ -1,6 +1,8 @@ -- -- MULTI_TASK_ASSIGNMENT -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 880000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 880000; SET citus.explain_distributed_queries TO off; -- Check that our policies for assigning tasks to worker nodes run as expected. -- To test this, we first create a shell table, and then manually insert shard diff --git a/src/test/regress/expected/multi_tpch_query1.out b/src/test/regress/expected/multi_tpch_query1.out index 23a507e06..6bc93dc4d 100644 --- a/src/test/regress/expected/multi_tpch_query1.out +++ b/src/test/regress/expected/multi_tpch_query1.out @@ -1,6 +1,8 @@ -- -- MULTI_TPCH_QUERY1 -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 890000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 890000; -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; -- Query #1 from the TPC-H decision support benchmark diff --git a/src/test/regress/expected/multi_tpch_query10.out b/src/test/regress/expected/multi_tpch_query10.out index 82eefa6df..2923dd432 100644 --- a/src/test/regress/expected/multi_tpch_query10.out +++ b/src/test/regress/expected/multi_tpch_query10.out @@ -4,6 +4,8 @@ -- Query #10 from the TPC-H decision support benchmark. Unlike other TPC-H tests, -- we don't set citus.large_table_shard_count here, and instead use the default value -- coming from postgresql.conf or multi_task_tracker_executor.conf. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 900000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 900000; SELECT c_custkey, c_name, diff --git a/src/test/regress/expected/multi_tpch_query12.out b/src/test/regress/expected/multi_tpch_query12.out index e54d3e0a4..ca9662c3e 100644 --- a/src/test/regress/expected/multi_tpch_query12.out +++ b/src/test/regress/expected/multi_tpch_query12.out @@ -1,6 +1,8 @@ -- -- MULTI_TPCH_QUERY12 -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 910000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 910000; -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; -- Query #12 from the TPC-H decision support benchmark diff --git a/src/test/regress/expected/multi_tpch_query14.out b/src/test/regress/expected/multi_tpch_query14.out index 2eded6dda..9bf8a5a82 100644 --- a/src/test/regress/expected/multi_tpch_query14.out +++ b/src/test/regress/expected/multi_tpch_query14.out @@ -1,6 +1,8 @@ -- -- MULTI_TPCH_QUERY14 -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 920000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 920000; -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; -- Query #14 from the TPC-H decision support benchmark diff --git a/src/test/regress/expected/multi_tpch_query19.out b/src/test/regress/expected/multi_tpch_query19.out index d6ee657c1..e0f1f0d86 100644 --- a/src/test/regress/expected/multi_tpch_query19.out +++ b/src/test/regress/expected/multi_tpch_query19.out @@ -1,6 +1,8 @@ -- -- MULTI_TPCH_QUERY19 -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 930000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 930000; -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; -- Query #19 from the TPC-H decision support benchmark. Note that we modified diff --git a/src/test/regress/expected/multi_tpch_query3.out b/src/test/regress/expected/multi_tpch_query3.out index da87dc8f5..67797e2e1 100644 --- a/src/test/regress/expected/multi_tpch_query3.out +++ b/src/test/regress/expected/multi_tpch_query3.out @@ -4,6 +4,8 @@ -- Query #3 from the TPC-H decision support benchmark. Unlike other TPC-H tests, -- we don't set citus.large_table_shard_count here, and instead use the default value -- coming from postgresql.conf or multi_task_tracker_executor.conf. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 940000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 940000; SELECT l_orderkey, sum(l_extendedprice * (1 - l_discount)) as revenue, diff --git a/src/test/regress/expected/multi_tpch_query6.out b/src/test/regress/expected/multi_tpch_query6.out index 9badebd9e..3d29efa07 100644 --- a/src/test/regress/expected/multi_tpch_query6.out +++ b/src/test/regress/expected/multi_tpch_query6.out @@ -1,6 +1,8 @@ -- -- MULTI_TPCH_QUERY6 -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 950000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 950000; -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; -- Query #6 from the TPC-H decision support benchmark diff --git a/src/test/regress/expected/multi_tpch_query7.out b/src/test/regress/expected/multi_tpch_query7.out index 55ab9ae63..69c8af9c4 100644 --- a/src/test/regress/expected/multi_tpch_query7.out +++ b/src/test/regress/expected/multi_tpch_query7.out @@ -1,6 +1,8 @@ -- -- MULTI_TPCH_QUERY7 -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 970000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 970000; -- Change configuration to treat lineitem AND orders tables as large SET citus.large_table_shard_count TO 2; -- Query #7 from the TPC-H decision support benchmark diff --git a/src/test/regress/expected/multi_tpch_query7_nested.out b/src/test/regress/expected/multi_tpch_query7_nested.out index d6d6acf71..1faa07b9c 100644 --- a/src/test/regress/expected/multi_tpch_query7_nested.out +++ b/src/test/regress/expected/multi_tpch_query7_nested.out @@ -1,6 +1,8 @@ -- -- MULTI_TPCH_QUERY7_NESTED -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 960000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 960000; -- Change configuration to treat lineitem AND orders tables AS large SET citus.large_table_shard_count TO 2; -- Query #7 from the TPC-H benchmark; modified to include sub-selects diff --git a/src/test/regress/expected/multi_upsert.out b/src/test/regress/expected/multi_upsert.out index 12864f4e4..ec9597171 100644 --- a/src/test/regress/expected/multi_upsert.out +++ b/src/test/regress/expected/multi_upsert.out @@ -1,6 +1,8 @@ -- this test file aims to test UPSERT feature on Citus -- note that output of this file for postgresql 9.4 will -- be full syntax errors, which is expected. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 980000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 980000; CREATE TABLE upsert_test ( part_key int UNIQUE, diff --git a/src/test/regress/expected/multi_upsert_0.out b/src/test/regress/expected/multi_upsert_0.out index c8c1e705e..09b052c1c 100644 --- a/src/test/regress/expected/multi_upsert_0.out +++ b/src/test/regress/expected/multi_upsert_0.out @@ -1,6 +1,8 @@ -- this test file aims to test UPSERT feature on Citus -- note that output of this file for postgresql 9.4 will -- be full syntax errors, which is expected. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 980000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 980000; CREATE TABLE upsert_test ( part_key int UNIQUE, diff --git a/src/test/regress/expected/multi_utilities.out b/src/test/regress/expected/multi_utilities.out index 6279c4806..ec20ec71b 100644 --- a/src/test/regress/expected/multi_utilities.out +++ b/src/test/regress/expected/multi_utilities.out @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 990000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 990000; -- =================================================================== -- test utility statement functionality -- =================================================================== diff --git a/src/test/regress/expected/multi_utility_statements.out b/src/test/regress/expected/multi_utility_statements.out index 1b502f20b..e4891e6c1 100644 --- a/src/test/regress/expected/multi_utility_statements.out +++ b/src/test/regress/expected/multi_utility_statements.out @@ -4,6 +4,8 @@ -- Check that we can run utility statements with embedded SELECT statements on -- distributed tables. Currently we only support CREATE TABLE AS (SELECT..), -- DECLARE CURSOR, and COPY ... TO statements. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1000000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1000000; CREATE TEMP TABLE lineitem_pricing_summary AS ( SELECT diff --git a/src/test/regress/expected/multi_utility_warnings.out b/src/test/regress/expected/multi_utility_warnings.out index b5ce4ec44..6d4f95406 100644 --- a/src/test/regress/expected/multi_utility_warnings.out +++ b/src/test/regress/expected/multi_utility_warnings.out @@ -3,6 +3,8 @@ -- -- Tests to check if we inform the user about potential caveats of creating new -- databases, schemas, and roles. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1010000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1010000; CREATE DATABASE new_database; NOTICE: Citus partially supports CREATE DATABASE for distributed databases DETAIL: Citus does not propagate CREATE DATABASE command to workers diff --git a/src/test/regress/expected/multi_verify_no_join_with_alias.out b/src/test/regress/expected/multi_verify_no_join_with_alias.out index 8d33250cd..0e6e0df20 100644 --- a/src/test/regress/expected/multi_verify_no_join_with_alias.out +++ b/src/test/regress/expected/multi_verify_no_join_with_alias.out @@ -3,6 +3,8 @@ -- -- This test checks that we simply emit an error message instead of trying to -- fetch and join a shard which has an alias set. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1020000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1020000; -- Show that the join works without an alias SELECT COUNT(*) FROM lineitem, part WHERE l_partkey = p_partkey; count @@ -11,10 +13,10 @@ SELECT COUNT(*) FROM lineitem, part WHERE l_partkey = p_partkey; (1 row) -- Assign an alias to the parts shard -UPDATE pg_dist_shard SET shardalias = 'my_alias' WHERE shardid = 102019; +UPDATE pg_dist_shard SET shardalias = 'my_alias' WHERE shardid = 290000; -- Attempt a join which uses this shard SELECT COUNT(*) FROM lineitem, part WHERE l_partkey = p_partkey; -ERROR: cannot fetch shard 102019 +ERROR: cannot fetch shard 290000 DETAIL: Fetching shards with aliases is currently unsupported -- Remove the alias from the parts shard -UPDATE pg_dist_shard SET shardalias = NULL WHERE shardid = 102019; +UPDATE pg_dist_shard SET shardalias = NULL WHERE shardid = 290000; diff --git a/src/test/regress/expected/multi_verify_no_subquery.out b/src/test/regress/expected/multi_verify_no_subquery.out index db1501098..74cef3a26 100644 --- a/src/test/regress/expected/multi_verify_no_subquery.out +++ b/src/test/regress/expected/multi_verify_no_subquery.out @@ -3,6 +3,8 @@ -- -- This test checks that we simply emit an error message instead of trying to -- process a distributed unsupported SQL subquery. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1030000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1030000; SELECT * FROM lineitem WHERE l_orderkey IN (SELECT l_orderkey FROM lineitem WHERE l_quantity > 0); ERROR: cannot perform distributed planning on this query diff --git a/src/test/regress/expected/multi_working_columns.out b/src/test/regress/expected/multi_working_columns.out index 38c05af49..13b5dadc1 100644 --- a/src/test/regress/expected/multi_working_columns.out +++ b/src/test/regress/expected/multi_working_columns.out @@ -5,6 +5,8 @@ -- projection order are called working (resjunk) columns. We check in here that -- these columns are pulled to the master, and are correctly used in sorting and -- grouping. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1040000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1040000; SELECT l_quantity FROM lineitem ORDER BY l_shipdate, l_quantity LIMIT 20; l_quantity ------------ diff --git a/src/test/regress/expected/task_tracker_assign_task.out b/src/test/regress/expected/task_tracker_assign_task.out index d9dff929b..ecbdfc461 100644 --- a/src/test/regress/expected/task_tracker_assign_task.out +++ b/src/test/regress/expected/task_tracker_assign_task.out @@ -1,6 +1,8 @@ -- -- TASK_TRACKER_ASSIGN_TASK -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1050000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1050000; \set JobId 401010 \set SimpleTaskId 101101 \set RecoverableTaskId 801102 diff --git a/src/test/regress/expected/task_tracker_cleanup_job.out b/src/test/regress/expected/task_tracker_cleanup_job.out index e1e63fbde..efe40e58b 100644 --- a/src/test/regress/expected/task_tracker_cleanup_job.out +++ b/src/test/regress/expected/task_tracker_cleanup_job.out @@ -1,6 +1,8 @@ -- -- TASK_TRACKER_CLEANUP_JOB -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1060000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1060000; \set JobId 401010 \set CompletedTaskId 801107 \set RunningTaskId 801108 diff --git a/src/test/regress/expected/task_tracker_create_table.out b/src/test/regress/expected/task_tracker_create_table.out index 30d45728c..d5fd46d63 100644 --- a/src/test/regress/expected/task_tracker_create_table.out +++ b/src/test/regress/expected/task_tracker_create_table.out @@ -1,6 +1,8 @@ -- -- TASK_TRACKER_CREATE_TABLE -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1070000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1070000; -- New table definitions to test the task tracker process and protocol CREATE TABLE lineitem_simple_task ( LIKE lineitem ); CREATE TABLE lineitem_compute_task ( LIKE lineitem ); diff --git a/src/test/regress/expected/task_tracker_partition_task.out b/src/test/regress/expected/task_tracker_partition_task.out index 2d36f93b0..bd61ada0f 100644 --- a/src/test/regress/expected/task_tracker_partition_task.out +++ b/src/test/regress/expected/task_tracker_partition_task.out @@ -1,6 +1,8 @@ -- -- TASK_TRACKER_PARTITION_TASK -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1080000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1080000; \set JobId 401010 \set PartitionTaskId 801106 \set PartitionColumn l_orderkey diff --git a/src/test/regress/expected/worker_binary_data_partition.out b/src/test/regress/expected/worker_binary_data_partition.out index afab11021..77d20f5e1 100644 --- a/src/test/regress/expected/worker_binary_data_partition.out +++ b/src/test/regress/expected/worker_binary_data_partition.out @@ -1,6 +1,8 @@ -- -- WORKER_BINARY_DATA_PARTITION -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1090000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1090000; \set JobId 201010 \set TaskId 101105 \set Partition_Column textcolumn diff --git a/src/test/regress/expected/worker_check_invalid_arguments.out b/src/test/regress/expected/worker_check_invalid_arguments.out index 1ecefdcb5..8affeac11 100644 --- a/src/test/regress/expected/worker_check_invalid_arguments.out +++ b/src/test/regress/expected/worker_check_invalid_arguments.out @@ -1,6 +1,8 @@ -- -- WORKER_CHECK_INVALID_ARGUMENTS -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1100000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1100000; \set JobId 201010 \set TaskId 101108 \set Table_Name simple_binary_data_table diff --git a/src/test/regress/expected/worker_create_table.out b/src/test/regress/expected/worker_create_table.out index 2d083a5ae..cc773b005 100644 --- a/src/test/regress/expected/worker_create_table.out +++ b/src/test/regress/expected/worker_create_table.out @@ -4,6 +4,8 @@ -- Create new table definitions for lineitem and supplier tables to test worker -- node execution logic. For now,the tests include range and hash partitioning -- of existing tables. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1110000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1110000; CREATE TABLE lineitem ( l_orderkey bigint not null, l_partkey integer not null, diff --git a/src/test/regress/expected/worker_hash_partition.out b/src/test/regress/expected/worker_hash_partition.out index f324571b2..3d7d5c94d 100644 --- a/src/test/regress/expected/worker_hash_partition.out +++ b/src/test/regress/expected/worker_hash_partition.out @@ -1,6 +1,8 @@ -- -- WORKER_HASH_PARTITION -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1130000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1130000; \set JobId 201010 \set TaskId 101103 \set Partition_Column l_orderkey diff --git a/src/test/regress/expected/worker_hash_partition_complex.out b/src/test/regress/expected/worker_hash_partition_complex.out index 2dfbe3702..29624a4c0 100644 --- a/src/test/regress/expected/worker_hash_partition_complex.out +++ b/src/test/regress/expected/worker_hash_partition_complex.out @@ -1,6 +1,8 @@ -- -- WORKER_HASH_PARTITION_COMPLEX -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1120000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1120000; \set JobId 201010 \set TaskId 101104 \set Partition_Column l_partkey diff --git a/src/test/regress/expected/worker_merge_hash_files.out b/src/test/regress/expected/worker_merge_hash_files.out index 89246b89f..671bbb5a5 100644 --- a/src/test/regress/expected/worker_merge_hash_files.out +++ b/src/test/regress/expected/worker_merge_hash_files.out @@ -1,6 +1,8 @@ -- -- WORKER_MERGE_HASH_FILES -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1140000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1140000; \set JobId 201010 \set TaskId 101103 \set Task_Table_Name public.task_101103 diff --git a/src/test/regress/expected/worker_merge_range_files.out b/src/test/regress/expected/worker_merge_range_files.out index b39f52731..264312af7 100644 --- a/src/test/regress/expected/worker_merge_range_files.out +++ b/src/test/regress/expected/worker_merge_range_files.out @@ -1,6 +1,8 @@ -- -- WORKER_MERGE_RANGE_FILES -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1150000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1150000; \set JobId 201010 \set TaskId 101101 \set Task_Table_Name public.task_101101 diff --git a/src/test/regress/expected/worker_null_data_partition.out b/src/test/regress/expected/worker_null_data_partition.out index 11881fbc9..9180fb1ec 100644 --- a/src/test/regress/expected/worker_null_data_partition.out +++ b/src/test/regress/expected/worker_null_data_partition.out @@ -1,6 +1,8 @@ -- -- WORKER_NULL_DATA_PARTITION -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1180000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1180000; \set JobId 201010 \set Range_TaskId 101106 \set Partition_Column s_nationkey diff --git a/src/test/regress/expected/worker_range_partition.out b/src/test/regress/expected/worker_range_partition.out index f33801018..85df6d48a 100644 --- a/src/test/regress/expected/worker_range_partition.out +++ b/src/test/regress/expected/worker_range_partition.out @@ -1,6 +1,8 @@ -- -- WORKER_RANGE_PARTITION -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1160000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1160000; \set JobId 201010 \set TaskId 101101 \set Partition_Column l_orderkey diff --git a/src/test/regress/expected/worker_range_partition_complex.out b/src/test/regress/expected/worker_range_partition_complex.out index 57b2c7795..fc68b2d32 100644 --- a/src/test/regress/expected/worker_range_partition_complex.out +++ b/src/test/regress/expected/worker_range_partition_complex.out @@ -1,6 +1,8 @@ -- -- WORKER_RANGE_PARTITION_COMPLEX -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1170000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1170000; \set JobId 201010 \set TaskId 101102 \set Partition_Column l_partkey diff --git a/src/test/regress/input/multi_agg_distinct.source b/src/test/regress/input/multi_agg_distinct.source index 469909b3d..10695bd9d 100644 --- a/src/test/regress/input/multi_agg_distinct.source +++ b/src/test/regress/input/multi_agg_distinct.source @@ -2,8 +2,12 @@ -- MULTI_AGG_DISTINCT -- --- Create a new range partitioned lineitem table and stage data into it +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 200000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 200000; + + +-- Create a new range partitioned lineitem table and stage data into it CREATE TABLE lineitem_range ( l_orderkey bigint not null, l_partkey integer not null, diff --git a/src/test/regress/input/multi_agg_type_conversion.source b/src/test/regress/input/multi_agg_type_conversion.source index 65e4375be..3573a2c15 100644 --- a/src/test/regress/input/multi_agg_type_conversion.source +++ b/src/test/regress/input/multi_agg_type_conversion.source @@ -2,8 +2,12 @@ -- MULTI_AGG_TYPE_CONVERSION -- --- Test aggregate type conversions using sums of integers and division operator +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 210000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 210000; + + +-- Test aggregate type conversions using sums of integers and division operator SELECT sum(l_suppkey) FROM lineitem; SELECT sum(l_suppkey) / 2 FROM lineitem; SELECT sum(l_suppkey) / 2::numeric FROM lineitem; diff --git a/src/test/regress/input/multi_alter_table_statements.source b/src/test/regress/input/multi_alter_table_statements.source index f7be6dc69..3df73f6db 100644 --- a/src/test/regress/input/multi_alter_table_statements.source +++ b/src/test/regress/input/multi_alter_table_statements.source @@ -2,11 +2,14 @@ -- MULTI_ALTER_TABLE_STATEMENTS -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 220000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 220000; + + -- Check that we can run ALTER TABLE statements on distributed tables. -- We set the shardid sequence here so that the shardids in this test -- aren't affected by changes to the previous tests. - -ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 103000; CREATE TABLE lineitem_alter ( l_orderkey bigint not null, l_partkey integer not null, @@ -178,7 +181,7 @@ ALTER TABLE lineitem_alter ADD COLUMN column_only_added_to_master int; -- verify newly added column is not present in a worker shard \c - - - :worker_1_port -SELECT column_only_added_to_master FROM lineitem_alter_103000 LIMIT 0; +SELECT column_only_added_to_master FROM lineitem_alter_220000 LIMIT 0; \c - - - :master_port -- ddl propagation flag is reset to default, disable it again diff --git a/src/test/regress/input/multi_append_table_to_shard.source b/src/test/regress/input/multi_append_table_to_shard.source index af0816f7d..4f453f9e0 100644 --- a/src/test/regress/input/multi_append_table_to_shard.source +++ b/src/test/regress/input/multi_append_table_to_shard.source @@ -1,6 +1,12 @@ -- -- MULTI_APPEND_TABLE_TO_SHARD -- + + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 230000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 230000; + + -- Initialize tables to join CREATE TABLE multi_append_table_to_shard_right ( diff --git a/src/test/regress/input/multi_complex_count_distinct.source b/src/test/regress/input/multi_complex_count_distinct.source index d9baa897a..7620e45a2 100644 --- a/src/test/regress/input/multi_complex_count_distinct.source +++ b/src/test/regress/input/multi_complex_count_distinct.source @@ -2,6 +2,11 @@ -- COMPLEX_COUNT_DISTINCT -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 240000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 240000; + + CREATE TABLE lineitem_hash ( l_orderkey bigint not null, l_partkey integer not null, diff --git a/src/test/regress/input/multi_copy.source b/src/test/regress/input/multi_copy.source index cf13682b8..daba9d182 100644 --- a/src/test/regress/input/multi_copy.source +++ b/src/test/regress/input/multi_copy.source @@ -2,7 +2,10 @@ -- MULTI_COPY -- + ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 560000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 560000; + -- Create a new hash-partitioned table into which to COPY CREATE TABLE customer_copy_hash ( diff --git a/src/test/regress/input/multi_create_schema.source b/src/test/regress/input/multi_create_schema.source index 333188d86..b2818f5ab 100644 --- a/src/test/regress/input/multi_create_schema.source +++ b/src/test/regress/input/multi_create_schema.source @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 250000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 250000; + + CREATE SCHEMA tpch CREATE TABLE nation ( n_nationkey integer not null, diff --git a/src/test/regress/input/multi_fdw_large_shardid.source b/src/test/regress/input/multi_fdw_large_shardid.source index 75a3b1cc5..cd5104c5e 100644 --- a/src/test/regress/input/multi_fdw_large_shardid.source +++ b/src/test/regress/input/multi_fdw_large_shardid.source @@ -6,7 +6,10 @@ -- differs from previous tests in that it modifies the *internal* shardId -- generator, forcing the distributed database to use 64-bit shard identifiers. + ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 100200300400500; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 9000000; + -- Stage additional data to start using large shard identifiers. diff --git a/src/test/regress/input/multi_fdw_stage_data.source b/src/test/regress/input/multi_fdw_stage_data.source index 8597407ca..0c7212b78 100644 --- a/src/test/regress/input/multi_fdw_stage_data.source +++ b/src/test/regress/input/multi_fdw_stage_data.source @@ -2,6 +2,11 @@ -- MULTI_FDW_STAGE_DATA -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 330000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 330000; + + -- Tests for staging foreign data in a distributed cluster. \STAGE lineitem FROM '@abs_srcdir@/data/lineitem.1.data' diff --git a/src/test/regress/input/multi_large_shardid.source b/src/test/regress/input/multi_large_shardid.source index 99f96a7df..21f69a209 100644 --- a/src/test/regress/input/multi_large_shardid.source +++ b/src/test/regress/input/multi_large_shardid.source @@ -6,7 +6,10 @@ -- differs from previous tests in that it modifies the *internal* shardId -- generator, forcing the distributed database to use 64-bit shard identifiers. + ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 100200300400500; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 9000000; + -- Stage additional data to start using large shard identifiers. diff --git a/src/test/regress/input/multi_master_delete_protocol.source b/src/test/regress/input/multi_master_delete_protocol.source index 646e0adf8..55ad68a3f 100644 --- a/src/test/regress/input/multi_master_delete_protocol.source +++ b/src/test/regress/input/multi_master_delete_protocol.source @@ -2,8 +2,12 @@ -- MULTI_MASTER_DELETE_PROTOCOL -- --- Create a new range partitioned customer_delete_protocol table and stage data into it. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 320000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 320000; + + +-- Create a new range partitioned customer_delete_protocol table and stage data into it. CREATE TABLE customer_delete_protocol ( c_custkey integer not null, c_name varchar(25) not null, diff --git a/src/test/regress/input/multi_outer_join.source b/src/test/regress/input/multi_outer_join.source index b0d0fce73..d8a6b11e1 100644 --- a/src/test/regress/input/multi_outer_join.source +++ b/src/test/regress/input/multi_outer_join.source @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 310000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 310000; + + SET citus.large_table_shard_count TO 2; SET citus.log_multi_join_order to true; SET client_min_messages TO LOG; diff --git a/src/test/regress/input/multi_stage_data.source b/src/test/regress/input/multi_stage_data.source index 9effe58d5..213521a98 100644 --- a/src/test/regress/input/multi_stage_data.source +++ b/src/test/regress/input/multi_stage_data.source @@ -1,12 +1,16 @@ -- -- MULTI_STAGE_DATA -- - -- Tests for staging data in a distributed cluster. Please note that the number -- of shards uploaded depends on two config values: citus.shard_replication_factor and -- citus.shard_max_size. These values are set in pg_regress_multi.pl. Shard placement -- policy is left to the default value (round-robin) to test the common install case. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 290000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 290000; + + \STAGE lineitem FROM '@abs_srcdir@/data/lineitem.1.data' with delimiter '|' \STAGE lineitem FROM '@abs_srcdir@/data/lineitem.2.data' with delimiter '|' diff --git a/src/test/regress/input/multi_stage_large_records.source b/src/test/regress/input/multi_stage_large_records.source index 8e015d537..496a6d757 100644 --- a/src/test/regress/input/multi_stage_large_records.source +++ b/src/test/regress/input/multi_stage_large_records.source @@ -1,11 +1,15 @@ -- -- MULTI_STAGE_LARGE_RECORDS -- - -- Tests for staging data with large records (i.e. greater than the read buffer -- size, which is 32kB) in a distributed cluster. These tests make sure that we -- are creating shards of correct size even when records are large. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 300000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 300000; + + SET citus.shard_max_size TO "256kB"; CREATE TABLE large_records_table (data_id integer, data text); diff --git a/src/test/regress/input/multi_stage_more_data.source b/src/test/regress/input/multi_stage_more_data.source index 3ec97813f..665c8c20d 100644 --- a/src/test/regress/input/multi_stage_more_data.source +++ b/src/test/regress/input/multi_stage_more_data.source @@ -2,6 +2,11 @@ -- MULTI_STAGE_MORE_DATA -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 280000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 280000; + + -- We stage more data to customer and part tables to test distributed joins. The -- staging causes the planner to consider customer and part tables as large, and -- evaluate plans where some of the underlying tables need to be repartitioned. diff --git a/src/test/regress/input/multi_subquery.source b/src/test/regress/input/multi_subquery.source index 032b044cf..52bbd7199 100644 --- a/src/test/regress/input/multi_subquery.source +++ b/src/test/regress/input/multi_subquery.source @@ -2,8 +2,12 @@ -- MULTI_SUBQUERY -- --- Create tables for subquery tests +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 270000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 270000; + + +-- Create tables for subquery tests CREATE TABLE lineitem_subquery ( l_orderkey bigint not null, l_partkey integer not null, @@ -101,7 +105,7 @@ FROM -- Update metadata in order to make all shards equal. -UPDATE pg_dist_shard SET shardmaxvalue = '14947' WHERE shardid = 102024; +UPDATE pg_dist_shard SET shardmaxvalue = '14947' WHERE shardid = 270003; -- If group by is not on partition column then we error out. @@ -350,5 +354,3 @@ SELECT * FROM AS foo; DROP TABLE subquery_pruning_varchar_test_table; - -ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 102026; diff --git a/src/test/regress/input/worker_copy.source b/src/test/regress/input/worker_copy.source index 0aba63e76..65ca778db 100644 --- a/src/test/regress/input/worker_copy.source +++ b/src/test/regress/input/worker_copy.source @@ -2,6 +2,11 @@ -- WORKER_COPY -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 260000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 260000; + + COPY lineitem FROM '@abs_srcdir@/data/lineitem.1.data' WITH DELIMITER '|'; COPY lineitem FROM '@abs_srcdir@/data/lineitem.2.data' WITH DELIMITER '|'; diff --git a/src/test/regress/output/multi_agg_distinct.source b/src/test/regress/output/multi_agg_distinct.source index d9a4967fa..42b25810f 100644 --- a/src/test/regress/output/multi_agg_distinct.source +++ b/src/test/regress/output/multi_agg_distinct.source @@ -1,6 +1,8 @@ -- -- MULTI_AGG_DISTINCT -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 200000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 200000; -- Create a new range partitioned lineitem table and stage data into it CREATE TABLE lineitem_range ( l_orderkey bigint not null, diff --git a/src/test/regress/output/multi_agg_type_conversion.source b/src/test/regress/output/multi_agg_type_conversion.source index cb73a98cb..49f65752f 100644 --- a/src/test/regress/output/multi_agg_type_conversion.source +++ b/src/test/regress/output/multi_agg_type_conversion.source @@ -1,6 +1,8 @@ -- -- MULTI_AGG_TYPE_CONVERSION -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 210000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 210000; -- Test aggregate type conversions using sums of integers and division operator SELECT sum(l_suppkey) FROM lineitem; sum diff --git a/src/test/regress/output/multi_alter_table_statements.source b/src/test/regress/output/multi_alter_table_statements.source index 378248546..467793f28 100644 --- a/src/test/regress/output/multi_alter_table_statements.source +++ b/src/test/regress/output/multi_alter_table_statements.source @@ -1,10 +1,11 @@ -- -- MULTI_ALTER_TABLE_STATEMENTS -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 220000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 220000; -- Check that we can run ALTER TABLE statements on distributed tables. -- We set the shardid sequence here so that the shardids in this test -- aren't affected by changes to the previous tests. -ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 103000; CREATE TABLE lineitem_alter ( l_orderkey bigint not null, l_partkey integer not null, @@ -166,10 +167,10 @@ ALTER TABLE lineitem_alter ALTER COLUMN int_column2 DROP DEFAULT; \STAGE lineitem_alter (l_orderkey, l_partkey, l_suppkey, l_linenumber, l_quantity, l_extendedprice, l_discount, l_tax, l_returnflag, l_linestatus, l_shipdate, l_commitdate, l_receiptdate, l_shipinstruct, l_shipmode, l_comment) FROM '@abs_srcdir@/data/lineitem.1.data' with delimiter '|' ERROR: null value in column "int_column2" violates not-null constraint DETAIL: Failing row contains (1, 155190, 7706, 1, 17.00, 21168.23, 0.04, 0.02, N, O, 03-13-1996, 02-12-1996, 03-22-1996, DELIVER IN PERSON , TRUCK , egular courts above the, 1, null, null, null, null). -CONTEXT: COPY lineitem_alter_103006, line 1: "1|155190|7706|1|17|21168.23|0.04|0.02|N|O|1996-03-13|1996-02-12|1996-03-22|DELIVER IN PERSON|TRUCK|e..." +CONTEXT: COPY lineitem_alter_220006, line 1: "1|155190|7706|1|17|21168.23|0.04|0.02|N|O|1996-03-13|1996-02-12|1996-03-22|DELIVER IN PERSON|TRUCK|e..." ERROR: null value in column "int_column2" violates not-null constraint DETAIL: Failing row contains (1, 155190, 7706, 1, 17.00, 21168.23, 0.04, 0.02, N, O, 03-13-1996, 02-12-1996, 03-22-1996, DELIVER IN PERSON , TRUCK , egular courts above the, 1, null, null, null, null). -CONTEXT: COPY lineitem_alter_103006, line 1: "1|155190|7706|1|17|21168.23|0.04|0.02|N|O|1996-03-13|1996-02-12|1996-03-22|DELIVER IN PERSON|TRUCK|e..." +CONTEXT: COPY lineitem_alter_220006, line 1: "1|155190|7706|1|17|21168.23|0.04|0.02|N|O|1996-03-13|1996-02-12|1996-03-22|DELIVER IN PERSON|TRUCK|e..." \stage: failed to replicate shard to enough replicas -- Verify that DROP NOT NULL works ALTER TABLE lineitem_alter ALTER COLUMN int_column2 DROP NOT NULL; @@ -261,7 +262,7 @@ NOTICE: relation "non_existent_table" does not exist, skipping ALTER TABLE IF EXISTS lineitem_alter ALTER COLUMN int_column2 SET DATA TYPE INTEGER; ALTER TABLE lineitem_alter DROP COLUMN non_existent_column; WARNING: Bad result from localhost:57638 -DETAIL: Remote message: column "non_existent_column" of relation "lineitem_alter_103000" does not exist +DETAIL: Remote message: column "non_existent_column" of relation "lineitem_alter_220000" does not exist ERROR: could not execute DDL command on worker node shards ALTER TABLE lineitem_alter DROP COLUMN IF EXISTS non_existent_column; NOTICE: column "non_existent_column" of relation "lineitem_alter" does not exist, skipping @@ -469,9 +470,9 @@ ALTER TABLE lineitem_renamed RENAME TO lineitem_alter; ALTER TABLE lineitem_alter ADD COLUMN column_only_added_to_master int; -- verify newly added column is not present in a worker shard \c - - - :worker_1_port -SELECT column_only_added_to_master FROM lineitem_alter_103000 LIMIT 0; +SELECT column_only_added_to_master FROM lineitem_alter_220000 LIMIT 0; ERROR: column "column_only_added_to_master" does not exist -LINE 1: SELECT column_only_added_to_master FROM lineitem_alter_10300... +LINE 1: SELECT column_only_added_to_master FROM lineitem_alter_22000... ^ \c - - - :master_port -- ddl propagation flag is reset to default, disable it again diff --git a/src/test/regress/output/multi_append_table_to_shard.source b/src/test/regress/output/multi_append_table_to_shard.source index e6f4fba5b..f3a3b3c91 100644 --- a/src/test/regress/output/multi_append_table_to_shard.source +++ b/src/test/regress/output/multi_append_table_to_shard.source @@ -1,6 +1,8 @@ -- -- MULTI_APPEND_TABLE_TO_SHARD -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 230000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 230000; -- Initialize tables to join CREATE TABLE multi_append_table_to_shard_right ( @@ -131,7 +133,7 @@ SELECT master_append_table_to_shard(shardid, 'multi_append_table_to_shard_stage' FROM pg_dist_shard WHERE 'multi_append_table_to_shard_right_hash'::regclass::oid = logicalrelid; -ERROR: cannot append to shardId 103011 +ERROR: cannot append to shardId 230000 DETAIL: We currently don't support appending to shards in hash-partitioned tables -- Clean up after test SELECT master_apply_delete_command('DELETE FROM multi_append_table_to_shard_right'); diff --git a/src/test/regress/output/multi_complex_count_distinct.source b/src/test/regress/output/multi_complex_count_distinct.source index 93b487834..b4c6b3a04 100644 --- a/src/test/regress/output/multi_complex_count_distinct.source +++ b/src/test/regress/output/multi_complex_count_distinct.source @@ -1,6 +1,8 @@ -- -- COMPLEX_COUNT_DISTINCT -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 240000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 240000; CREATE TABLE lineitem_hash ( l_orderkey bigint not null, l_partkey integer not null, diff --git a/src/test/regress/output/multi_copy.source b/src/test/regress/output/multi_copy.source index 8a5d49a31..ff3990b46 100644 --- a/src/test/regress/output/multi_copy.source +++ b/src/test/regress/output/multi_copy.source @@ -2,6 +2,7 @@ -- MULTI_COPY -- ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 560000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 560000; -- Create a new hash-partitioned table into which to COPY CREATE TABLE customer_copy_hash ( c_custkey integer, diff --git a/src/test/regress/output/multi_create_schema.source b/src/test/regress/output/multi_create_schema.source index 690a309d4..0ef71898a 100644 --- a/src/test/regress/output/multi_create_schema.source +++ b/src/test/regress/output/multi_create_schema.source @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 250000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 250000; CREATE SCHEMA tpch CREATE TABLE nation ( n_nationkey integer not null, diff --git a/src/test/regress/output/multi_fdw_large_shardid.source b/src/test/regress/output/multi_fdw_large_shardid.source index 485280e68..fac7dd83f 100644 --- a/src/test/regress/output/multi_fdw_large_shardid.source +++ b/src/test/regress/output/multi_fdw_large_shardid.source @@ -5,6 +5,7 @@ -- differs from previous tests in that it modifies the *internal* shardId -- generator, forcing the distributed database to use 64-bit shard identifiers. ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 100200300400500; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 9000000; -- Stage additional data to start using large shard identifiers. \STAGE lineitem FROM '@abs_srcdir@/data/lineitem.1.data' NOTICE: extension "file_fdw" already exists, skipping diff --git a/src/test/regress/output/multi_fdw_stage_data.source b/src/test/regress/output/multi_fdw_stage_data.source index 3cb781624..0605a9404 100644 --- a/src/test/regress/output/multi_fdw_stage_data.source +++ b/src/test/regress/output/multi_fdw_stage_data.source @@ -1,6 +1,8 @@ -- -- MULTI_FDW_STAGE_DATA -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 330000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 330000; -- Tests for staging foreign data in a distributed cluster. \STAGE lineitem FROM '@abs_srcdir@/data/lineitem.1.data' NOTICE: extension "file_fdw" already exists, skipping diff --git a/src/test/regress/output/multi_large_shardid.source b/src/test/regress/output/multi_large_shardid.source index 65d8f6bda..684732ee6 100644 --- a/src/test/regress/output/multi_large_shardid.source +++ b/src/test/regress/output/multi_large_shardid.source @@ -5,6 +5,7 @@ -- differs from previous tests in that it modifies the *internal* shardId -- generator, forcing the distributed database to use 64-bit shard identifiers. ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 100200300400500; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 9000000; -- Stage additional data to start using large shard identifiers. \STAGE lineitem FROM '@abs_srcdir@/data/lineitem.1.data' with delimiter '|' \STAGE lineitem FROM '@abs_srcdir@/data/lineitem.2.data' with delimiter '|' diff --git a/src/test/regress/output/multi_master_delete_protocol.source b/src/test/regress/output/multi_master_delete_protocol.source index a1de0751c..15c9ee819 100644 --- a/src/test/regress/output/multi_master_delete_protocol.source +++ b/src/test/regress/output/multi_master_delete_protocol.source @@ -1,6 +1,8 @@ -- -- MULTI_MASTER_DELETE_PROTOCOL -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 320000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 320000; -- Create a new range partitioned customer_delete_protocol table and stage data into it. CREATE TABLE customer_delete_protocol ( c_custkey integer not null, diff --git a/src/test/regress/output/multi_outer_join.source b/src/test/regress/output/multi_outer_join.source index ac7b7eb3c..0abc638bc 100644 --- a/src/test/regress/output/multi_outer_join.source +++ b/src/test/regress/output/multi_outer_join.source @@ -1,3 +1,5 @@ +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 310000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 310000; SET citus.large_table_shard_count TO 2; SET citus.log_multi_join_order to true; SET client_min_messages TO LOG; @@ -301,6 +303,16 @@ FROM LOG: join order: [ "multi_outer_join_left" ][ local partition join "multi_outer_join_right" ][ broadcast join "multi_outer_join_third" ] l_custkey | r_custkey | t_custkey -----------+-----------+----------- + 21 | 21 | 21 + 22 | 22 | 22 + 23 | 23 | 23 + 24 | 24 | 24 + 25 | 25 | 25 + 26 | 26 | 26 + 27 | 27 | 27 + 28 | 28 | 28 + 29 | 29 | 29 + 30 | 30 | 30 1 | | 2 | | 3 | | @@ -316,16 +328,6 @@ LOG: join order: [ "multi_outer_join_left" ][ local partition join "multi_outer 13 | 13 | 13 14 | 14 | 14 15 | 15 | 15 - 21 | 21 | 21 - 22 | 22 | 22 - 23 | 23 | 23 - 24 | 24 | 24 - 25 | 25 | 25 - 26 | 26 | 26 - 27 | 27 | 27 - 28 | 28 | 28 - 29 | 29 | 29 - 30 | 30 | 30 (25 rows) -- Right join with single shard right most table should error out @@ -347,16 +349,6 @@ FROM LOG: join order: [ "multi_outer_join_right" ][ broadcast join "multi_outer_join_third" ][ local partition join "multi_outer_join_left" ] t_custkey | r_custkey | l_custkey -----------+-----------+----------- - 11 | 11 | 11 - 12 | 12 | 12 - 13 | 13 | 13 - 14 | 14 | 14 - 15 | 15 | 15 - 16 | 16 | - 17 | 17 | - 18 | 18 | - 19 | 19 | - 20 | 20 | 21 | 21 | 21 22 | 22 | 22 23 | 23 | 23 @@ -367,6 +359,16 @@ LOG: join order: [ "multi_outer_join_right" ][ broadcast join "multi_outer_join 28 | 28 | 28 29 | 29 | 29 30 | 30 | 30 + 11 | 11 | 11 + 12 | 12 | 12 + 13 | 13 | 13 + 14 | 14 | 14 + 15 | 15 | 15 + 16 | 16 | + 17 | 17 | + 18 | 18 | + 19 | 19 | + 20 | 20 | (20 rows) -- Make it anti-join, should display values with l_custkey is null @@ -406,6 +408,16 @@ FROM LOG: join order: [ "multi_outer_join_left" ][ local partition join "multi_outer_join_right" ] l_custkey | r_custkey -----------+----------- + 21 | 21 + 22 | 22 + 23 | 23 + 24 | 24 + 25 | 25 + 26 | 26 + 27 | 27 + 28 | 28 + 29 | 29 + 30 | 30 1 | 2 | 3 | @@ -426,16 +438,6 @@ LOG: join order: [ "multi_outer_join_left" ][ local partition join "multi_outer | 18 | 19 | 16 - 21 | 21 - 22 | 22 - 23 | 23 - 24 | 24 - 25 | 25 - 26 | 26 - 27 | 27 - 28 | 28 - 29 | 29 - 30 | 30 (30 rows) -- full outer join + anti (right) should work with 1-1 matched shards @@ -525,11 +527,6 @@ FROM LOG: join order: [ "multi_outer_join_left" ][ local partition join "multi_outer_join_right" ][ broadcast join "multi_outer_join_third" ] l_custkey | r_custkey | t_custkey -----------+-----------+----------- - 11 | 11 | 11 - 12 | 12 | 12 - 13 | 13 | 13 - 14 | 14 | 14 - 15 | 15 | 15 21 | 21 | 21 22 | 22 | 22 23 | 23 | 23 @@ -540,6 +537,11 @@ LOG: join order: [ "multi_outer_join_left" ][ local partition join "multi_outer 28 | 28 | 28 29 | 29 | 29 30 | 30 | 30 + 11 | 11 | 11 + 12 | 12 | 12 + 13 | 13 | 13 + 14 | 14 | 14 + 15 | 15 | 15 (15 rows) -- inner (broadcast) join + 2 shards left (local) join should work @@ -552,6 +554,16 @@ FROM LOG: join order: [ "multi_outer_join_left" ][ broadcast join "multi_outer_join_third" ][ local partition join "multi_outer_join_right" ] l_custkey | t_custkey | r_custkey -----------+-----------+----------- + 21 | 21 | 21 + 22 | 22 | 22 + 23 | 23 | 23 + 24 | 24 | 24 + 25 | 25 | 25 + 26 | 26 | 26 + 27 | 27 | 27 + 28 | 28 | 28 + 29 | 29 | 29 + 30 | 30 | 30 1 | 1 | 2 | 2 | 3 | 3 | @@ -567,16 +579,6 @@ LOG: join order: [ "multi_outer_join_left" ][ broadcast join "multi_outer_join_ 13 | 13 | 13 14 | 14 | 14 15 | 15 | 15 - 21 | 21 | 21 - 22 | 22 | 22 - 23 | 23 | 23 - 24 | 24 | 24 - 25 | 25 | 25 - 26 | 26 | 26 - 27 | 27 | 27 - 28 | 28 | 28 - 29 | 29 | 29 - 30 | 30 | 30 (25 rows) -- inner (local) join + 2 shards left (dual partition) join should error out @@ -598,6 +600,16 @@ FROM LOG: join order: [ "multi_outer_join_left" ][ broadcast join "multi_outer_join_third" ][ local partition join "multi_outer_join_right" ] l_custkey | t_custkey | r_custkey -----------+-----------+----------- + 21 | 21 | 21 + 22 | 22 | 22 + 23 | 23 | 23 + 24 | 24 | 24 + 25 | 25 | 25 + 26 | 26 | 26 + 27 | 27 | 27 + 28 | 28 | 28 + 29 | 29 | 29 + 30 | 30 | 30 1 | 1 | 2 | 2 | 3 | 3 | @@ -613,16 +625,6 @@ LOG: join order: [ "multi_outer_join_left" ][ broadcast join "multi_outer_join_ 13 | 13 | 13 14 | 14 | 14 15 | 15 | 15 - 21 | 21 | 21 - 22 | 22 | 22 - 23 | 23 | 23 - 24 | 24 | 24 - 25 | 25 | 25 - 26 | 26 | 26 - 27 | 27 | 27 - 28 | 28 | 28 - 29 | 29 | 29 - 30 | 30 | 30 (25 rows) -- inner (broadcast) join + 2 shards left (local) + anti join should work @@ -660,16 +662,6 @@ FROM LOG: join order: [ "multi_outer_join_right" ][ local partition join "multi_outer_join_left" ][ broadcast join "multi_outer_join_third" ] t_custkey ----------- - 11 - 12 - 13 - 14 - 15 - 16 - 17 - 18 - 19 - 20 21 22 23 @@ -680,6 +672,16 @@ LOG: join order: [ "multi_outer_join_right" ][ local partition join "multi_oute 28 29 30 + 11 + 12 + 13 + 14 + 15 + 16 + 17 + 18 + 19 + 20 (20 rows) -- Add a shard to the left table that overlaps with multiple shards in the right diff --git a/src/test/regress/output/multi_stage_data.source b/src/test/regress/output/multi_stage_data.source index 0eed8965f..b322feb76 100644 --- a/src/test/regress/output/multi_stage_data.source +++ b/src/test/regress/output/multi_stage_data.source @@ -5,6 +5,8 @@ -- of shards uploaded depends on two config values: citus.shard_replication_factor and -- citus.shard_max_size. These values are set in pg_regress_multi.pl. Shard placement -- policy is left to the default value (round-robin) to test the common install case. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 290000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 290000; \STAGE lineitem FROM '@abs_srcdir@/data/lineitem.1.data' with delimiter '|' \STAGE lineitem FROM '@abs_srcdir@/data/lineitem.2.data' with delimiter '|' \STAGE orders FROM '@abs_srcdir@/data/orders.1.data' with delimiter '|' diff --git a/src/test/regress/output/multi_stage_large_records.source b/src/test/regress/output/multi_stage_large_records.source index 24eaa14e6..ef707850c 100644 --- a/src/test/regress/output/multi_stage_large_records.source +++ b/src/test/regress/output/multi_stage_large_records.source @@ -4,6 +4,8 @@ -- Tests for staging data with large records (i.e. greater than the read buffer -- size, which is 32kB) in a distributed cluster. These tests make sure that we -- are creating shards of correct size even when records are large. +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 300000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 300000; SET citus.shard_max_size TO "256kB"; CREATE TABLE large_records_table (data_id integer, data text); SELECT master_create_distributed_table('large_records_table', 'data_id', 'append'); diff --git a/src/test/regress/output/multi_stage_more_data.source b/src/test/regress/output/multi_stage_more_data.source index 2e0f38f1d..3c2c5eba5 100644 --- a/src/test/regress/output/multi_stage_more_data.source +++ b/src/test/regress/output/multi_stage_more_data.source @@ -1,6 +1,8 @@ -- -- MULTI_STAGE_MORE_DATA -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 280000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 280000; -- We stage more data to customer and part tables to test distributed joins. The -- staging causes the planner to consider customer and part tables as large, and -- evaluate plans where some of the underlying tables need to be repartitioned. diff --git a/src/test/regress/output/multi_subquery.source b/src/test/regress/output/multi_subquery.source index e6724b673..15ee33f2e 100644 --- a/src/test/regress/output/multi_subquery.source +++ b/src/test/regress/output/multi_subquery.source @@ -1,6 +1,8 @@ -- -- MULTI_SUBQUERY -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 270000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 270000; -- Create tables for subquery tests CREATE TABLE lineitem_subquery ( l_orderkey bigint not null, @@ -103,7 +105,7 @@ FROM ERROR: cannot push down this subquery DETAIL: Shards of relations in subquery need to have 1-to-1 shard partitioning -- Update metadata in order to make all shards equal. -UPDATE pg_dist_shard SET shardmaxvalue = '14947' WHERE shardid = 102024; +UPDATE pg_dist_shard SET shardmaxvalue = '14947' WHERE shardid = 270003; -- If group by is not on partition column then we error out. SELECT avg(order_count) @@ -349,9 +351,9 @@ SET client_min_messages TO DEBUG2; SELECT * FROM (SELECT count(*) FROM subquery_pruning_varchar_test_table WHERE a = 'onder' GROUP BY a) AS foo; -DEBUG: predicate pruning for shardId 102026 -DEBUG: predicate pruning for shardId 102027 -DEBUG: predicate pruning for shardId 102029 +DEBUG: predicate pruning for shardId 270005 +DEBUG: predicate pruning for shardId 270006 +DEBUG: predicate pruning for shardId 270008 count ------- (0 rows) @@ -359,9 +361,9 @@ DEBUG: predicate pruning for shardId 102029 SELECT * FROM (SELECT count(*) FROM subquery_pruning_varchar_test_table WHERE 'eren' = a GROUP BY a) AS foo; -DEBUG: predicate pruning for shardId 102026 -DEBUG: predicate pruning for shardId 102028 -DEBUG: predicate pruning for shardId 102029 +DEBUG: predicate pruning for shardId 270005 +DEBUG: predicate pruning for shardId 270007 +DEBUG: predicate pruning for shardId 270008 count ------- (0 rows) @@ -400,4 +402,3 @@ AS foo; (0 rows) DROP TABLE subquery_pruning_varchar_test_table; -ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 102026; diff --git a/src/test/regress/output/worker_copy.source b/src/test/regress/output/worker_copy.source index 954d47931..31c3c19c4 100644 --- a/src/test/regress/output/worker_copy.source +++ b/src/test/regress/output/worker_copy.source @@ -1,6 +1,8 @@ -- -- WORKER_COPY -- +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 260000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 260000; COPY lineitem FROM '@abs_srcdir@/data/lineitem.1.data' WITH DELIMITER '|'; COPY lineitem FROM '@abs_srcdir@/data/lineitem.2.data' WITH DELIMITER '|'; COPY supplier FROM '@abs_srcdir@/data/supplier.data' WITH DELIMITER '|'; diff --git a/src/test/regress/sql/multi_agg_approximate_distinct.sql b/src/test/regress/sql/multi_agg_approximate_distinct.sql index 5ed4d527c..a79be18a1 100644 --- a/src/test/regress/sql/multi_agg_approximate_distinct.sql +++ b/src/test/regress/sql/multi_agg_approximate_distinct.sql @@ -2,6 +2,11 @@ -- MULTI_AGG_APPROXIMATE_DISTINCT -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 340000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 340000; + + -- Try to execute count(distinct) when approximate distincts aren't enabled SELECT count(distinct l_orderkey) FROM lineitem; diff --git a/src/test/regress/sql/multi_array_agg.sql b/src/test/regress/sql/multi_array_agg.sql index d4031827a..541d5ebd9 100644 --- a/src/test/regress/sql/multi_array_agg.sql +++ b/src/test/regress/sql/multi_array_agg.sql @@ -2,6 +2,11 @@ -- MULTI_ARRAY_AGG -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 520000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 520000; + + -- Check multi_cat_agg() aggregate which is used to implement array_agg() SELECT array_cat_agg(i) FROM (VALUES (ARRAY[1,2]), (NULL), (ARRAY[3,4])) AS t(i); diff --git a/src/test/regress/sql/multi_average_expression.sql b/src/test/regress/sql/multi_average_expression.sql index 856dfc74c..68d07625e 100644 --- a/src/test/regress/sql/multi_average_expression.sql +++ b/src/test/regress/sql/multi_average_expression.sql @@ -1,11 +1,15 @@ -- -- MULTI_AVERAGE_EXPRESSION_ORDER -- - -- This test checks that the group-by columns don't need to be above an average -- expression, and can be anywhere in the projection order. This is in response -- to a bug we had due to the average expression introducing new columns. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 450000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 450000; + + SELECT sum(l_quantity) as sum_qty, sum(l_extendedprice) as sum_base_price, diff --git a/src/test/regress/sql/multi_basic_queries.sql b/src/test/regress/sql/multi_basic_queries.sql index 10e195725..f00d0a148 100644 --- a/src/test/regress/sql/multi_basic_queries.sql +++ b/src/test/regress/sql/multi_basic_queries.sql @@ -2,6 +2,11 @@ -- MULTI_BASIC_QUERIES -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 440000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 440000; + + -- Execute simple sum, average, and count queries on data recently uploaded to -- our partitioned table. diff --git a/src/test/regress/sql/multi_binary_master_copy_format.sql b/src/test/regress/sql/multi_binary_master_copy_format.sql index 4c67eb2d2..94ca3ccfb 100644 --- a/src/test/regress/sql/multi_binary_master_copy_format.sql +++ b/src/test/regress/sql/multi_binary_master_copy_format.sql @@ -2,6 +2,11 @@ -- MULTI_BINARY_MASTER_COPY -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 430000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 430000; + + -- Try binary master copy for different executors SET citus.binary_master_copy_format TO 'on'; diff --git a/src/test/regress/sql/multi_complex_expressions.sql b/src/test/regress/sql/multi_complex_expressions.sql index 1eaafc86c..89205be85 100644 --- a/src/test/regress/sql/multi_complex_expressions.sql +++ b/src/test/regress/sql/multi_complex_expressions.sql @@ -2,6 +2,11 @@ -- MULTI_COMPLEX_EXPRESSIONS -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 420000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 420000; + + -- Check that we can correctly handle complex expressions and aggregates. SELECT sum(l_quantity) / avg(l_quantity) FROM lineitem; diff --git a/src/test/regress/sql/multi_connection_cache.sql b/src/test/regress/sql/multi_connection_cache.sql index ea1bdff7d..442545c73 100644 --- a/src/test/regress/sql/multi_connection_cache.sql +++ b/src/test/regress/sql/multi_connection_cache.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 410000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 410000; + + -- =================================================================== -- create test functions -- =================================================================== diff --git a/src/test/regress/sql/multi_count_type_conversion.sql b/src/test/regress/sql/multi_count_type_conversion.sql index 2f84da988..ee45ff11d 100644 --- a/src/test/regress/sql/multi_count_type_conversion.sql +++ b/src/test/regress/sql/multi_count_type_conversion.sql @@ -2,6 +2,11 @@ -- MULTI_COUNT_TYPE_CONVERSION -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 400000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 400000; + + -- Verify that we can sort count(*) results correctly. We perform this check as -- our count() operations execute in two steps: worker nodes report their -- count() results, and the master node sums these counts up. During this sum(), diff --git a/src/test/regress/sql/multi_create_fdw.sql b/src/test/regress/sql/multi_create_fdw.sql index dd828a66b..e8bd25386 100644 --- a/src/test/regress/sql/multi_create_fdw.sql +++ b/src/test/regress/sql/multi_create_fdw.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 390000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 390000; + + -- =================================================================== -- get ready for the foreign data wrapper tests -- =================================================================== diff --git a/src/test/regress/sql/multi_create_insert_proxy.sql b/src/test/regress/sql/multi_create_insert_proxy.sql index bfd3adcbd..ee8de14d8 100644 --- a/src/test/regress/sql/multi_create_insert_proxy.sql +++ b/src/test/regress/sql/multi_create_insert_proxy.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 380000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 380000; + + -- =================================================================== -- test INSERT proxy creation functionality -- =================================================================== diff --git a/src/test/regress/sql/multi_create_shards.sql b/src/test/regress/sql/multi_create_shards.sql index 1180338ac..c5e38fa63 100644 --- a/src/test/regress/sql/multi_create_shards.sql +++ b/src/test/regress/sql/multi_create_shards.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 370000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 370000; + + -- =================================================================== -- create test functions and types needed for tests -- =================================================================== diff --git a/src/test/regress/sql/multi_create_table.sql b/src/test/regress/sql/multi_create_table.sql index 3509436f6..4d9db7e37 100644 --- a/src/test/regress/sql/multi_create_table.sql +++ b/src/test/regress/sql/multi_create_table.sql @@ -2,6 +2,11 @@ -- MULTI_CREATE_TABLE -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 360000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 360000; + + -- Create new table definitions for use in testing in distributed planning and -- execution functionality. Also create indexes to boost performance. diff --git a/src/test/regress/sql/multi_data_types.sql b/src/test/regress/sql/multi_data_types.sql index 68b84044f..eaf4c41b4 100644 --- a/src/test/regress/sql/multi_data_types.sql +++ b/src/test/regress/sql/multi_data_types.sql @@ -3,6 +3,11 @@ -- create, distribute, INSERT, SELECT and UPDATE -- =================================================================== + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 530000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 530000; + + -- create a custom type... CREATE TYPE test_composite_type AS ( i integer, diff --git a/src/test/regress/sql/multi_distribution_metadata.sql b/src/test/regress/sql/multi_distribution_metadata.sql index 897c44c71..01c0a2c76 100644 --- a/src/test/regress/sql/multi_distribution_metadata.sql +++ b/src/test/regress/sql/multi_distribution_metadata.sql @@ -2,6 +2,11 @@ -- create test functions -- =================================================================== + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 540000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 540000; + + CREATE FUNCTION load_shard_id_array(regclass) RETURNS bigint[] AS 'citus' @@ -82,13 +87,13 @@ SELECT master_create_distributed_table('events_hash', 'name', 'hash'); SELECT master_create_worker_shards('events_hash', 4, 2); -- set shardstate of one replication from each shard to 0 (invalid value) -UPDATE pg_dist_shard_placement SET shardstate = 0 WHERE nodeport = 57638 AND shardid BETWEEN 103025 AND 103028; +UPDATE pg_dist_shard_placement SET shardstate = 0 WHERE nodeport = 57638 AND shardid BETWEEN 540000 AND 540003; -- should see above shard identifiers SELECT load_shard_id_array('events_hash'); -- should see array with first shard range -SELECT load_shard_interval_array(103025, 0); +SELECT load_shard_interval_array(540000, 0); -- should even work for range-partitioned shards -- create range distributed table @@ -104,21 +109,21 @@ SELECT master_create_empty_shard('events_range'); UPDATE pg_dist_shard SET shardminvalue = 'Aardvark', shardmaxvalue = 'Zebra' -WHERE shardid = 103029; +WHERE shardid = 540004; -SELECT load_shard_interval_array(103029, ''::text); +SELECT load_shard_interval_array(540004, ''::text); -- should see error for non-existent shard -SELECT load_shard_interval_array(103030, 0); +SELECT load_shard_interval_array(540005, 0); -- should see two placements -SELECT load_shard_placement_array(103026, false); +SELECT load_shard_placement_array(540001, false); -- only one of which is finalized -SELECT load_shard_placement_array(103026, true); +SELECT load_shard_placement_array(540001, true); -- should see error for non-existent shard -SELECT load_shard_placement_array(103031, false); +SELECT load_shard_placement_array(540001, false); -- should see column id of 'name' SELECT partition_column_id('events_hash'); @@ -139,7 +144,7 @@ SELECT column_name_to_column_id('events_hash', 'non_existent'); -- drop shard rows (must drop placements first) DELETE FROM pg_dist_shard_placement - WHERE shardid BETWEEN 103025 AND 103029; + WHERE shardid BETWEEN 540000 AND 540004; DELETE FROM pg_dist_shard WHERE logicalrelid = 'events_hash'::regclass; DELETE FROM pg_dist_shard diff --git a/src/test/regress/sql/multi_drop_extension.sql b/src/test/regress/sql/multi_drop_extension.sql index 41ee4f896..c5320260d 100644 --- a/src/test/regress/sql/multi_drop_extension.sql +++ b/src/test/regress/sql/multi_drop_extension.sql @@ -4,6 +4,10 @@ -- Tests around dropping and recreating the extension +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 550000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 550000; + + CREATE TABLE testtableddl(somecol int, distributecol text NOT NULL); SELECT master_create_distributed_table('testtableddl', 'distributecol', 'append'); diff --git a/src/test/regress/sql/multi_dropped_column_aliases.sql b/src/test/regress/sql/multi_dropped_column_aliases.sql index 689e64a00..3c791ff28 100644 --- a/src/test/regress/sql/multi_dropped_column_aliases.sql +++ b/src/test/regress/sql/multi_dropped_column_aliases.sql @@ -2,6 +2,11 @@ -- Tests that check that our query functionality behaves as expected when the -- table schema is modified via ALTER statements. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 620000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 620000; + + SELECT count(*) FROM customer; SELECT * FROM customer LIMIT 2; diff --git a/src/test/regress/sql/multi_explain.sql b/src/test/regress/sql/multi_explain.sql index 3138a03d7..aa16e9d3a 100644 --- a/src/test/regress/sql/multi_explain.sql +++ b/src/test/regress/sql/multi_explain.sql @@ -2,6 +2,11 @@ -- MULTI_EXPLAIN -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 570000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 570000; + + \a\t SET citus.task_executor_type TO 'real-time'; diff --git a/src/test/regress/sql/multi_extension.sql b/src/test/regress/sql/multi_extension.sql index 0d33a0b82..24701b91a 100644 --- a/src/test/regress/sql/multi_extension.sql +++ b/src/test/regress/sql/multi_extension.sql @@ -6,6 +6,11 @@ -- It'd be nice to script generation of this file, but alas, that's -- not done yet. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 580000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 580000; + + -- DROP EXTENSION pre-created by the regression suite DROP EXTENSION citus; \c diff --git a/src/test/regress/sql/multi_fdw_create_table.sql b/src/test/regress/sql/multi_fdw_create_table.sql index e0ccd9a3e..89b9342cc 100644 --- a/src/test/regress/sql/multi_fdw_create_table.sql +++ b/src/test/regress/sql/multi_fdw_create_table.sql @@ -2,6 +2,11 @@ -- MULTI_FDW_CREATE_TABLE -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 590000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 590000; + + -- Create new table definitions for use in testing in distributed foreign data -- wrapper functionality. diff --git a/src/test/regress/sql/multi_fdw_master_protocol.sql b/src/test/regress/sql/multi_fdw_master_protocol.sql index dfd495538..7efb1ab61 100644 --- a/src/test/regress/sql/multi_fdw_master_protocol.sql +++ b/src/test/regress/sql/multi_fdw_master_protocol.sql @@ -4,6 +4,11 @@ -- Tests that check the metadata returned by the master node. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 600000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 600000; + + SELECT part_storage_type, part_key, part_replica_count, part_max_size, part_placement_policy FROM master_get_table_metadata('lineitem'); diff --git a/src/test/regress/sql/multi_generate_ddl_commands.sql b/src/test/regress/sql/multi_generate_ddl_commands.sql index 0bd2c28fa..3e7b644a3 100644 --- a/src/test/regress/sql/multi_generate_ddl_commands.sql +++ b/src/test/regress/sql/multi_generate_ddl_commands.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 610000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 610000; + + -- =================================================================== -- create test functions -- =================================================================== diff --git a/src/test/regress/sql/multi_hash_pruning.sql b/src/test/regress/sql/multi_hash_pruning.sql index 981b6b22c..d6210d8f8 100644 --- a/src/test/regress/sql/multi_hash_pruning.sql +++ b/src/test/regress/sql/multi_hash_pruning.sql @@ -4,6 +4,11 @@ -- Tests for shard and join pruning logic on hash partitioned tables. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 630000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 630000; + + -- Create a table partitioned on integer column and update partition type to -- hash. Then stage data to this table and update shard min max values with -- hashed ones. Hash value of 1, 2, 3 and 4 are consecutively -1905060026, diff --git a/src/test/regress/sql/multi_index_statements.sql b/src/test/regress/sql/multi_index_statements.sql index 66be10f63..a71d5b82d 100644 --- a/src/test/regress/sql/multi_index_statements.sql +++ b/src/test/regress/sql/multi_index_statements.sql @@ -5,6 +5,11 @@ -- Check that we can run CREATE INDEX and DROP INDEX statements on distributed -- tables. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 640000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 640000; + + -- -- CREATE TEST TABLES -- diff --git a/src/test/regress/sql/multi_join_order_additional.sql b/src/test/regress/sql/multi_join_order_additional.sql index 221a62538..fd650e6f9 100644 --- a/src/test/regress/sql/multi_join_order_additional.sql +++ b/src/test/regress/sql/multi_join_order_additional.sql @@ -2,6 +2,11 @@ -- MULTI_JOIN_ORDER_ADDITIONAL -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 650000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 650000; + + -- Set configuration to print table join order and pruned shards SET citus.explain_distributed_queries TO off; diff --git a/src/test/regress/sql/multi_join_order_tpch_large.sql b/src/test/regress/sql/multi_join_order_tpch_large.sql index ed8ca08bd..89ab65411 100644 --- a/src/test/regress/sql/multi_join_order_tpch_large.sql +++ b/src/test/regress/sql/multi_join_order_tpch_large.sql @@ -2,6 +2,11 @@ -- MULTI_JOIN_ORDER_TPCH_LARGE -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 660000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 660000; + + -- Enable configuration to print table join order SET citus.explain_distributed_queries TO off; diff --git a/src/test/regress/sql/multi_join_pruning.sql b/src/test/regress/sql/multi_join_pruning.sql index f43632b5f..17c9bc6e1 100644 --- a/src/test/regress/sql/multi_join_pruning.sql +++ b/src/test/regress/sql/multi_join_pruning.sql @@ -2,6 +2,10 @@ -- MULTI_JOIN_PRUNING -- + +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 680000; + + -- Check that join-pruning works for joins between two large relations. For now -- we only check for join-pruning between locally partitioned relations. In the -- future we want to check for pruning between re-partitioned relations as well. diff --git a/src/test/regress/sql/multi_large_table_join_planning.sql b/src/test/regress/sql/multi_large_table_join_planning.sql index 8ad2a3fe6..91c2d25e6 100644 --- a/src/test/regress/sql/multi_large_table_join_planning.sql +++ b/src/test/regress/sql/multi_large_table_join_planning.sql @@ -1,12 +1,16 @@ -- -- MULTI_LARGE_TABLE_PLANNING -- - -- Tests that cover large table join planning. Note that we explicitly start a -- transaction block here so that we don't emit debug messages with changing -- transaction ids in them. Also, we set the executor type to task tracker -- executor here, as we cannot run repartition jobs with real time executor. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 690000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 690000; + + BEGIN; SET client_min_messages TO DEBUG4; SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_large_table_pruning.sql b/src/test/regress/sql/multi_large_table_pruning.sql index ec46eff0c..5a1f80c8c 100644 --- a/src/test/regress/sql/multi_large_table_pruning.sql +++ b/src/test/regress/sql/multi_large_table_pruning.sql @@ -1,11 +1,15 @@ -- -- MULTI_LARGE_TABLE_PRUNING -- - -- Tests covering partition and join-pruning for large table joins. Note that we -- set executor type to task tracker executor here, as we cannot run repartition -- jobs with real time executor. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 700000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 700000; + + SET citus.large_table_shard_count TO 2; SET client_min_messages TO DEBUG2; SET citus.task_executor_type TO 'task-tracker'; diff --git a/src/test/regress/sql/multi_large_table_task_assignment.sql b/src/test/regress/sql/multi_large_table_task_assignment.sql index 0793e4492..bc4d0fbb7 100644 --- a/src/test/regress/sql/multi_large_table_task_assignment.sql +++ b/src/test/regress/sql/multi_large_table_task_assignment.sql @@ -1,12 +1,16 @@ -- -- MULTI_LARGE_TABLE_TASK_ASSIGNMENT -- - -- Tests which cover task assignment for MapMerge jobs for single range repartition -- and dual hash repartition joins. The tests also cover task assignment propagation -- from a sql task to its depended tasks. Note that we set the executor type to task -- tracker executor here, as we cannot run repartition jobs with real time executor. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 710000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 710000; + + BEGIN; SET client_min_messages TO DEBUG3; SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_limit_clause.sql b/src/test/regress/sql/multi_limit_clause.sql index 9d68ec526..7cf7a1578 100644 --- a/src/test/regress/sql/multi_limit_clause.sql +++ b/src/test/regress/sql/multi_limit_clause.sql @@ -2,6 +2,11 @@ -- MULTI_LIMIT_CLAUSE -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 730000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 730000; + + -- Display debug messages on limit clause push down. SET client_min_messages TO DEBUG1; diff --git a/src/test/regress/sql/multi_limit_clause_approximate.sql b/src/test/regress/sql/multi_limit_clause_approximate.sql index 659e91ab6..1e82a1c44 100644 --- a/src/test/regress/sql/multi_limit_clause_approximate.sql +++ b/src/test/regress/sql/multi_limit_clause_approximate.sql @@ -2,6 +2,11 @@ -- MULTI_LIMIT_CLAUSE_APPROXIMATE -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 720000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 720000; + + -- Display debug messages on limit clause push down. SET client_min_messages TO DEBUG1; diff --git a/src/test/regress/sql/multi_master_protocol.sql b/src/test/regress/sql/multi_master_protocol.sql index 054220e76..2440619bb 100644 --- a/src/test/regress/sql/multi_master_protocol.sql +++ b/src/test/regress/sql/multi_master_protocol.sql @@ -1,9 +1,13 @@ -- -- MULTI_MASTER_PROTOCOL -- - -- Tests that check the metadata returned by the master node. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 740000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 740000; + + SELECT part_storage_type, part_key, part_replica_count, part_max_size, part_placement_policy FROM master_get_table_metadata('lineitem'); diff --git a/src/test/regress/sql/multi_modifications.sql b/src/test/regress/sql/multi_modifications.sql index 69444d372..238fa57d2 100644 --- a/src/test/regress/sql/multi_modifications.sql +++ b/src/test/regress/sql/multi_modifications.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 750000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 750000; + + -- =================================================================== -- test end-to-end modification functionality -- =================================================================== diff --git a/src/test/regress/sql/multi_null_minmax_value_pruning.sql b/src/test/regress/sql/multi_null_minmax_value_pruning.sql index b471e926e..db8ce464b 100644 --- a/src/test/regress/sql/multi_null_minmax_value_pruning.sql +++ b/src/test/regress/sql/multi_null_minmax_value_pruning.sql @@ -5,14 +5,19 @@ -- This test checks that we can handle null min/max values in shard statistics -- and that we don't partition or join prune shards that have null values. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 760000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 760000; + + SET client_min_messages TO DEBUG2; -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; -SELECT shardminvalue, shardmaxvalue from pg_dist_shard WHERE shardid = 102009; -SELECT shardminvalue, shardmaxvalue from pg_dist_shard WHERE shardid = 102010; +SELECT shardminvalue, shardmaxvalue from pg_dist_shard WHERE shardid = 290000; +SELECT shardminvalue, shardmaxvalue from pg_dist_shard WHERE shardid = 290001; -- Check that partition and join pruning works when min/max values exist -- Adding l_orderkey = 1 to make the query not router executable @@ -24,7 +29,7 @@ SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem, orders -- Now set the minimum value for a shard to null. Then check that we don't apply -- partition or join pruning for the shard with null min value. -UPDATE pg_dist_shard SET shardminvalue = NULL WHERE shardid = 102009; +UPDATE pg_dist_shard SET shardminvalue = NULL WHERE shardid = 290000; SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 9030; @@ -34,7 +39,7 @@ SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem, orders -- Next, set the maximum value for another shard to null. Then check that we -- don't apply partition or join pruning for this other shard either. -UPDATE pg_dist_shard SET shardmaxvalue = NULL WHERE shardid = 102010; +UPDATE pg_dist_shard SET shardmaxvalue = NULL WHERE shardid = 290001; SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 9030; @@ -44,7 +49,7 @@ SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem, orders -- Last, set the minimum value to 0 and check that we don't treat it as null. We -- should apply partition and join pruning for this shard now. -UPDATE pg_dist_shard SET shardminvalue = '0' WHERE shardid = 102009; +UPDATE pg_dist_shard SET shardminvalue = '0' WHERE shardid = 290000; SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 9030; @@ -53,7 +58,7 @@ SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem, orders -- Set minimum and maximum values for two shards back to their original values -UPDATE pg_dist_shard SET shardminvalue = '1' WHERE shardid = 102009; -UPDATE pg_dist_shard SET shardmaxvalue = '4964' WHERE shardid = 102010; +UPDATE pg_dist_shard SET shardminvalue = '1' WHERE shardid = 290000; +UPDATE pg_dist_shard SET shardmaxvalue = '4964' WHERE shardid = 290001; SET client_min_messages TO NOTICE; diff --git a/src/test/regress/sql/multi_partition_pruning.sql b/src/test/regress/sql/multi_partition_pruning.sql index 128ec266a..18e58d5f5 100644 --- a/src/test/regress/sql/multi_partition_pruning.sql +++ b/src/test/regress/sql/multi_partition_pruning.sql @@ -5,6 +5,11 @@ -- Tests to verify that we correctly prune unreferenced shards. For this, we -- need to increase the logging verbosity of messages displayed on the client. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 770000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 770000; + + SET citus.explain_distributed_queries TO off; SET client_min_messages TO DEBUG2; diff --git a/src/test/regress/sql/multi_prepare_plsql.sql b/src/test/regress/sql/multi_prepare_plsql.sql index f62e256df..8fd1d0bd4 100644 --- a/src/test/regress/sql/multi_prepare_plsql.sql +++ b/src/test/regress/sql/multi_prepare_plsql.sql @@ -6,6 +6,11 @@ -- and converted into both plain SQL and PL/pgsql functions, which -- use prepared statements internally. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 780000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 780000; + + CREATE FUNCTION sql_test_no_1() RETURNS bigint AS ' SELECT count(*) diff --git a/src/test/regress/sql/multi_prepare_sql.sql b/src/test/regress/sql/multi_prepare_sql.sql index 346b8de6a..18417729c 100644 --- a/src/test/regress/sql/multi_prepare_sql.sql +++ b/src/test/regress/sql/multi_prepare_sql.sql @@ -6,6 +6,11 @@ -- taken from other regression test files and converted into -- prepared statements. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 790000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 790000; + + PREPARE prepared_test_1 AS SELECT count(*) diff --git a/src/test/regress/sql/multi_prune_shard_list.sql b/src/test/regress/sql/multi_prune_shard_list.sql index 25d7be0fc..9dc60213e 100644 --- a/src/test/regress/sql/multi_prune_shard_list.sql +++ b/src/test/regress/sql/multi_prune_shard_list.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 800000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 800000; + + -- =================================================================== -- create test functions -- =================================================================== diff --git a/src/test/regress/sql/multi_query_directory_cleanup.sql b/src/test/regress/sql/multi_query_directory_cleanup.sql index 0f83e4802..4192495e9 100644 --- a/src/test/regress/sql/multi_query_directory_cleanup.sql +++ b/src/test/regress/sql/multi_query_directory_cleanup.sql @@ -7,6 +7,11 @@ -- the resource owner should automatically clean up these intermediate query -- result files. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 810000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 810000; + + BEGIN; -- pg_ls_dir() displays jobids. We explicitly set the jobId sequence diff --git a/src/test/regress/sql/multi_repair_shards.sql b/src/test/regress/sql/multi_repair_shards.sql index 7912f6a34..e7647a350 100644 --- a/src/test/regress/sql/multi_repair_shards.sql +++ b/src/test/regress/sql/multi_repair_shards.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 820000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 820000; + + -- =================================================================== -- test shard repair functionality -- =================================================================== diff --git a/src/test/regress/sql/multi_repartitioned_subquery_udf.sql b/src/test/regress/sql/multi_repartitioned_subquery_udf.sql index a245cd420..59d8ea030 100644 --- a/src/test/regress/sql/multi_repartitioned_subquery_udf.sql +++ b/src/test/regress/sql/multi_repartitioned_subquery_udf.sql @@ -2,6 +2,11 @@ -- MULTI_REPARTITIONED_SUBQUERY_UDF -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 830000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 830000; + + -- Create UDF in master and workers \c - - - :master_port DROP FUNCTION IF EXISTS median(double precision[]); diff --git a/src/test/regress/sql/multi_router_planner.sql b/src/test/regress/sql/multi_router_planner.sql index 7bbc9a985..a7da7efde 100644 --- a/src/test/regress/sql/multi_router_planner.sql +++ b/src/test/regress/sql/multi_router_planner.sql @@ -1,8 +1,11 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 840000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 840000; + + -- =================================================================== -- test router planner functionality for single shard select queries -- =================================================================== -ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 103300; - CREATE TABLE articles_hash ( id bigint NOT NULL, diff --git a/src/test/regress/sql/multi_shard_modify.sql b/src/test/regress/sql/multi_shard_modify.sql index 76c511889..65d28a7d3 100644 --- a/src/test/regress/sql/multi_shard_modify.sql +++ b/src/test/regress/sql/multi_shard_modify.sql @@ -2,7 +2,10 @@ -- MULTI_SHARD_MODIFY -- + ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 350000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 350000; + -- Create a new hash partitioned multi_shard_modify_test table and stage data into it. CREATE TABLE multi_shard_modify_test ( diff --git a/src/test/regress/sql/multi_simple_queries.sql b/src/test/regress/sql/multi_simple_queries.sql index a2ccca852..0f479738a 100644 --- a/src/test/regress/sql/multi_simple_queries.sql +++ b/src/test/regress/sql/multi_simple_queries.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 850000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 850000; + + -- =================================================================== -- test end-to-end query functionality -- =================================================================== diff --git a/src/test/regress/sql/multi_single_relation_subquery.sql b/src/test/regress/sql/multi_single_relation_subquery.sql index a170af39a..688de921e 100644 --- a/src/test/regress/sql/multi_single_relation_subquery.sql +++ b/src/test/regress/sql/multi_single_relation_subquery.sql @@ -1,9 +1,13 @@ -- -- MULTI_SINGLE_RELATION_SUBQUERY -- - -- This test checks that we are able to run selected set of distributed SQL subqueries. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 860000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 860000; + + SET citus.task_executor_type TO 'task-tracker'; select diff --git a/src/test/regress/sql/multi_table_ddl.sql b/src/test/regress/sql/multi_table_ddl.sql index 51247bf89..c79703777 100644 --- a/src/test/regress/sql/multi_table_ddl.sql +++ b/src/test/regress/sql/multi_table_ddl.sql @@ -3,6 +3,11 @@ -- -- Tests around changing the schema and dropping of a distributed table + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 870000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 870000; + + CREATE TABLE testtableddl(somecol int, distributecol text NOT NULL); SELECT master_create_distributed_table('testtableddl', 'distributecol', 'append'); diff --git a/src/test/regress/sql/multi_task_assignment_policy.sql b/src/test/regress/sql/multi_task_assignment_policy.sql index 1fd30b716..72dcdff11 100644 --- a/src/test/regress/sql/multi_task_assignment_policy.sql +++ b/src/test/regress/sql/multi_task_assignment_policy.sql @@ -2,8 +2,13 @@ -- MULTI_TASK_ASSIGNMENT -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 880000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 880000; + SET citus.explain_distributed_queries TO off; + -- Check that our policies for assigning tasks to worker nodes run as expected. -- To test this, we first create a shell table, and then manually insert shard -- and shard placement data into system catalogs. We next run Explain command, diff --git a/src/test/regress/sql/multi_tpch_query1.sql b/src/test/regress/sql/multi_tpch_query1.sql index d6727d15a..6a2f9f5e4 100644 --- a/src/test/regress/sql/multi_tpch_query1.sql +++ b/src/test/regress/sql/multi_tpch_query1.sql @@ -2,6 +2,11 @@ -- MULTI_TPCH_QUERY1 -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 890000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 890000; + + -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_tpch_query10.sql b/src/test/regress/sql/multi_tpch_query10.sql index 75bb475a5..9f136f267 100644 --- a/src/test/regress/sql/multi_tpch_query10.sql +++ b/src/test/regress/sql/multi_tpch_query10.sql @@ -6,6 +6,11 @@ -- we don't set citus.large_table_shard_count here, and instead use the default value -- coming from postgresql.conf or multi_task_tracker_executor.conf. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 900000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 900000; + + SELECT c_custkey, c_name, diff --git a/src/test/regress/sql/multi_tpch_query12.sql b/src/test/regress/sql/multi_tpch_query12.sql index 380165bb2..170dbae7d 100644 --- a/src/test/regress/sql/multi_tpch_query12.sql +++ b/src/test/regress/sql/multi_tpch_query12.sql @@ -2,6 +2,11 @@ -- MULTI_TPCH_QUERY12 -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 910000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 910000; + + -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_tpch_query14.sql b/src/test/regress/sql/multi_tpch_query14.sql index 82b696fec..5a730e95b 100644 --- a/src/test/regress/sql/multi_tpch_query14.sql +++ b/src/test/regress/sql/multi_tpch_query14.sql @@ -2,6 +2,11 @@ -- MULTI_TPCH_QUERY14 -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 920000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 920000; + + -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_tpch_query19.sql b/src/test/regress/sql/multi_tpch_query19.sql index dfff2ea5d..52fc0e5b8 100644 --- a/src/test/regress/sql/multi_tpch_query19.sql +++ b/src/test/regress/sql/multi_tpch_query19.sql @@ -2,6 +2,11 @@ -- MULTI_TPCH_QUERY19 -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 930000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 930000; + + -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_tpch_query3.sql b/src/test/regress/sql/multi_tpch_query3.sql index 62232e5fa..f96768b99 100644 --- a/src/test/regress/sql/multi_tpch_query3.sql +++ b/src/test/regress/sql/multi_tpch_query3.sql @@ -6,6 +6,11 @@ -- we don't set citus.large_table_shard_count here, and instead use the default value -- coming from postgresql.conf or multi_task_tracker_executor.conf. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 940000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 940000; + + SELECT l_orderkey, sum(l_extendedprice * (1 - l_discount)) as revenue, diff --git a/src/test/regress/sql/multi_tpch_query6.sql b/src/test/regress/sql/multi_tpch_query6.sql index 994b9467b..af48ecb51 100644 --- a/src/test/regress/sql/multi_tpch_query6.sql +++ b/src/test/regress/sql/multi_tpch_query6.sql @@ -2,6 +2,11 @@ -- MULTI_TPCH_QUERY6 -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 950000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 950000; + + -- Change configuration to treat lineitem and orders tables as large SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_tpch_query7.sql b/src/test/regress/sql/multi_tpch_query7.sql index b9b3c5cd3..f0287200e 100644 --- a/src/test/regress/sql/multi_tpch_query7.sql +++ b/src/test/regress/sql/multi_tpch_query7.sql @@ -2,6 +2,11 @@ -- MULTI_TPCH_QUERY7 -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 970000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 970000; + + -- Change configuration to treat lineitem AND orders tables as large SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_tpch_query7_nested.sql b/src/test/regress/sql/multi_tpch_query7_nested.sql index ad34dd7b5..7a80a93a2 100644 --- a/src/test/regress/sql/multi_tpch_query7_nested.sql +++ b/src/test/regress/sql/multi_tpch_query7_nested.sql @@ -2,6 +2,11 @@ -- MULTI_TPCH_QUERY7_NESTED -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 960000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 960000; + + -- Change configuration to treat lineitem AND orders tables AS large SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_upsert.sql b/src/test/regress/sql/multi_upsert.sql index b61833eaf..394ef0bbc 100644 --- a/src/test/regress/sql/multi_upsert.sql +++ b/src/test/regress/sql/multi_upsert.sql @@ -2,6 +2,11 @@ -- note that output of this file for postgresql 9.4 will -- be full syntax errors, which is expected. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 980000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 980000; + + CREATE TABLE upsert_test ( part_key int UNIQUE, diff --git a/src/test/regress/sql/multi_utilities.sql b/src/test/regress/sql/multi_utilities.sql index 7070311f7..15fa16cdc 100644 --- a/src/test/regress/sql/multi_utilities.sql +++ b/src/test/regress/sql/multi_utilities.sql @@ -1,3 +1,8 @@ + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 990000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 990000; + + -- =================================================================== -- test utility statement functionality -- =================================================================== diff --git a/src/test/regress/sql/multi_utility_statements.sql b/src/test/regress/sql/multi_utility_statements.sql index e3ebdcec6..a29842208 100644 --- a/src/test/regress/sql/multi_utility_statements.sql +++ b/src/test/regress/sql/multi_utility_statements.sql @@ -6,6 +6,11 @@ -- distributed tables. Currently we only support CREATE TABLE AS (SELECT..), -- DECLARE CURSOR, and COPY ... TO statements. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1000000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1000000; + + CREATE TEMP TABLE lineitem_pricing_summary AS ( SELECT diff --git a/src/test/regress/sql/multi_utility_warnings.sql b/src/test/regress/sql/multi_utility_warnings.sql index 6f7503fa4..c53b9ac5b 100644 --- a/src/test/regress/sql/multi_utility_warnings.sql +++ b/src/test/regress/sql/multi_utility_warnings.sql @@ -5,6 +5,11 @@ -- Tests to check if we inform the user about potential caveats of creating new -- databases, schemas, and roles. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1010000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1010000; + + CREATE DATABASE new_database; CREATE SCHEMA new_schema; diff --git a/src/test/regress/sql/multi_verify_no_join_with_alias.sql b/src/test/regress/sql/multi_verify_no_join_with_alias.sql index b957d6227..18e446df1 100644 --- a/src/test/regress/sql/multi_verify_no_join_with_alias.sql +++ b/src/test/regress/sql/multi_verify_no_join_with_alias.sql @@ -5,13 +5,18 @@ -- This test checks that we simply emit an error message instead of trying to -- fetch and join a shard which has an alias set. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1020000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1020000; + + -- Show that the join works without an alias SELECT COUNT(*) FROM lineitem, part WHERE l_partkey = p_partkey; -- Assign an alias to the parts shard -UPDATE pg_dist_shard SET shardalias = 'my_alias' WHERE shardid = 102019; +UPDATE pg_dist_shard SET shardalias = 'my_alias' WHERE shardid = 290000; -- Attempt a join which uses this shard @@ -19,4 +24,4 @@ SELECT COUNT(*) FROM lineitem, part WHERE l_partkey = p_partkey; -- Remove the alias from the parts shard -UPDATE pg_dist_shard SET shardalias = NULL WHERE shardid = 102019; +UPDATE pg_dist_shard SET shardalias = NULL WHERE shardid = 290000; diff --git a/src/test/regress/sql/multi_verify_no_subquery.sql b/src/test/regress/sql/multi_verify_no_subquery.sql index d6d1c1b41..b94eddcca 100644 --- a/src/test/regress/sql/multi_verify_no_subquery.sql +++ b/src/test/regress/sql/multi_verify_no_subquery.sql @@ -5,6 +5,11 @@ -- This test checks that we simply emit an error message instead of trying to -- process a distributed unsupported SQL subquery. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1030000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1030000; + + SELECT * FROM lineitem WHERE l_orderkey IN (SELECT l_orderkey FROM lineitem WHERE l_quantity > 0); diff --git a/src/test/regress/sql/multi_working_columns.sql b/src/test/regress/sql/multi_working_columns.sql index ae792ae2b..aaee2dcb3 100644 --- a/src/test/regress/sql/multi_working_columns.sql +++ b/src/test/regress/sql/multi_working_columns.sql @@ -7,6 +7,11 @@ -- these columns are pulled to the master, and are correctly used in sorting and -- grouping. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1040000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1040000; + + SELECT l_quantity FROM lineitem ORDER BY l_shipdate, l_quantity LIMIT 20; SELECT l_quantity, count(*) as count FROM lineitem diff --git a/src/test/regress/sql/task_tracker_assign_task.sql b/src/test/regress/sql/task_tracker_assign_task.sql index d9e24ee65..34bfd56f7 100644 --- a/src/test/regress/sql/task_tracker_assign_task.sql +++ b/src/test/regress/sql/task_tracker_assign_task.sql @@ -2,6 +2,11 @@ -- TASK_TRACKER_ASSIGN_TASK -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1050000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1050000; + + \set JobId 401010 \set SimpleTaskId 101101 \set RecoverableTaskId 801102 diff --git a/src/test/regress/sql/task_tracker_cleanup_job.sql b/src/test/regress/sql/task_tracker_cleanup_job.sql index be026d62f..2717f5ddf 100644 --- a/src/test/regress/sql/task_tracker_cleanup_job.sql +++ b/src/test/regress/sql/task_tracker_cleanup_job.sql @@ -2,6 +2,11 @@ -- TASK_TRACKER_CLEANUP_JOB -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1060000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1060000; + + \set JobId 401010 \set CompletedTaskId 801107 \set RunningTaskId 801108 diff --git a/src/test/regress/sql/task_tracker_create_table.sql b/src/test/regress/sql/task_tracker_create_table.sql index d27e3354e..322d8fe23 100644 --- a/src/test/regress/sql/task_tracker_create_table.sql +++ b/src/test/regress/sql/task_tracker_create_table.sql @@ -2,6 +2,11 @@ -- TASK_TRACKER_CREATE_TABLE -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1070000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1070000; + + -- New table definitions to test the task tracker process and protocol CREATE TABLE lineitem_simple_task ( LIKE lineitem ); diff --git a/src/test/regress/sql/task_tracker_partition_task.sql b/src/test/regress/sql/task_tracker_partition_task.sql index a77092882..0f465f823 100644 --- a/src/test/regress/sql/task_tracker_partition_task.sql +++ b/src/test/regress/sql/task_tracker_partition_task.sql @@ -2,6 +2,11 @@ -- TASK_TRACKER_PARTITION_TASK -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1080000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1080000; + + \set JobId 401010 \set PartitionTaskId 801106 diff --git a/src/test/regress/sql/worker_binary_data_partition.sql b/src/test/regress/sql/worker_binary_data_partition.sql index 0c2d487dc..1094dff89 100644 --- a/src/test/regress/sql/worker_binary_data_partition.sql +++ b/src/test/regress/sql/worker_binary_data_partition.sql @@ -2,6 +2,11 @@ -- WORKER_BINARY_DATA_PARTITION -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1090000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1090000; + + \set JobId 201010 \set TaskId 101105 \set Partition_Column textcolumn diff --git a/src/test/regress/sql/worker_check_invalid_arguments.sql b/src/test/regress/sql/worker_check_invalid_arguments.sql index 464e92985..6feb68ddc 100644 --- a/src/test/regress/sql/worker_check_invalid_arguments.sql +++ b/src/test/regress/sql/worker_check_invalid_arguments.sql @@ -2,6 +2,11 @@ -- WORKER_CHECK_INVALID_ARGUMENTS -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1100000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1100000; + + \set JobId 201010 \set TaskId 101108 \set Table_Name simple_binary_data_table diff --git a/src/test/regress/sql/worker_create_table.sql b/src/test/regress/sql/worker_create_table.sql index 7f9699d65..6ba2fe462 100644 --- a/src/test/regress/sql/worker_create_table.sql +++ b/src/test/regress/sql/worker_create_table.sql @@ -6,6 +6,11 @@ -- node execution logic. For now,the tests include range and hash partitioning -- of existing tables. + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1110000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1110000; + + CREATE TABLE lineitem ( l_orderkey bigint not null, l_partkey integer not null, diff --git a/src/test/regress/sql/worker_hash_partition.sql b/src/test/regress/sql/worker_hash_partition.sql index 0bf5e2ef8..ff1b7ec32 100644 --- a/src/test/regress/sql/worker_hash_partition.sql +++ b/src/test/regress/sql/worker_hash_partition.sql @@ -2,6 +2,11 @@ -- WORKER_HASH_PARTITION -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1130000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1130000; + + \set JobId 201010 \set TaskId 101103 \set Partition_Column l_orderkey diff --git a/src/test/regress/sql/worker_hash_partition_complex.sql b/src/test/regress/sql/worker_hash_partition_complex.sql index 0aa89ef79..32d1810fa 100644 --- a/src/test/regress/sql/worker_hash_partition_complex.sql +++ b/src/test/regress/sql/worker_hash_partition_complex.sql @@ -2,6 +2,11 @@ -- WORKER_HASH_PARTITION_COMPLEX -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1120000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1120000; + + \set JobId 201010 \set TaskId 101104 \set Partition_Column l_partkey diff --git a/src/test/regress/sql/worker_merge_hash_files.sql b/src/test/regress/sql/worker_merge_hash_files.sql index 3f8c59f00..bba66cf37 100644 --- a/src/test/regress/sql/worker_merge_hash_files.sql +++ b/src/test/regress/sql/worker_merge_hash_files.sql @@ -2,6 +2,11 @@ -- WORKER_MERGE_HASH_FILES -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1140000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1140000; + + \set JobId 201010 \set TaskId 101103 \set Task_Table_Name public.task_101103 diff --git a/src/test/regress/sql/worker_merge_range_files.sql b/src/test/regress/sql/worker_merge_range_files.sql index 26f67e94b..cd7189258 100644 --- a/src/test/regress/sql/worker_merge_range_files.sql +++ b/src/test/regress/sql/worker_merge_range_files.sql @@ -2,6 +2,11 @@ -- WORKER_MERGE_RANGE_FILES -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1150000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1150000; + + \set JobId 201010 \set TaskId 101101 \set Task_Table_Name public.task_101101 diff --git a/src/test/regress/sql/worker_null_data_partition.sql b/src/test/regress/sql/worker_null_data_partition.sql index b696c8d14..ba9aa06ed 100644 --- a/src/test/regress/sql/worker_null_data_partition.sql +++ b/src/test/regress/sql/worker_null_data_partition.sql @@ -2,6 +2,11 @@ -- WORKER_NULL_DATA_PARTITION -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1180000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1180000; + + \set JobId 201010 \set Range_TaskId 101106 \set Partition_Column s_nationkey diff --git a/src/test/regress/sql/worker_range_partition.sql b/src/test/regress/sql/worker_range_partition.sql index f6c0d87b2..5d8d4a4f8 100644 --- a/src/test/regress/sql/worker_range_partition.sql +++ b/src/test/regress/sql/worker_range_partition.sql @@ -2,6 +2,11 @@ -- WORKER_RANGE_PARTITION -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1160000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1160000; + + \set JobId 201010 \set TaskId 101101 \set Partition_Column l_orderkey diff --git a/src/test/regress/sql/worker_range_partition_complex.sql b/src/test/regress/sql/worker_range_partition_complex.sql index 14a536a57..109cc95e1 100644 --- a/src/test/regress/sql/worker_range_partition_complex.sql +++ b/src/test/regress/sql/worker_range_partition_complex.sql @@ -2,6 +2,11 @@ -- WORKER_RANGE_PARTITION_COMPLEX -- + +ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 1170000; +ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 1170000; + + \set JobId 201010 \set TaskId 101102 \set Partition_Column l_partkey