ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 990000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 990000; -- =================================================================== -- test utility statement functionality -- =================================================================== CREATE TABLE sharded_table ( name text, id bigint ); SELECT master_create_distributed_table('sharded_table', 'id', 'hash'); master_create_distributed_table --------------------------------- (1 row) SELECT master_create_worker_shards('sharded_table', 2, 1); master_create_worker_shards ----------------------------- (1 row) -- COPY out is supported with distributed tables COPY sharded_table TO STDOUT; COPY (SELECT COUNT(*) FROM sharded_table) TO STDOUT; 0 -- cursors may not involve distributed tables DECLARE all_sharded_rows CURSOR FOR SELECT * FROM sharded_table; ERROR: DECLARE CURSOR can only be used in transaction blocks -- verify PREPARE functionality PREPARE sharded_insert AS INSERT INTO sharded_table VALUES ('adam', 1); PREPARE sharded_update AS UPDATE sharded_table SET name = 'bob' WHERE id = 1; PREPARE sharded_delete AS DELETE FROM sharded_table WHERE id = 1; PREPARE sharded_query AS SELECT name FROM sharded_table WHERE id = 1; EXECUTE sharded_query; name ------ (0 rows) EXECUTE sharded_insert; EXECUTE sharded_query; name ------ adam (1 row) EXECUTE sharded_update; EXECUTE sharded_query; name ------ bob (1 row) EXECUTE sharded_delete; EXECUTE sharded_query; name ------ (0 rows) -- try to drop shards with where clause SELECT master_apply_delete_command('DELETE FROM sharded_table WHERE id > 0'); ERROR: cannot delete from distributed table DETAIL: Delete statements on hash-partitioned tables with where clause is not supported -- drop all shards SELECT master_apply_delete_command('DELETE FROM sharded_table'); master_apply_delete_command ----------------------------- 2 (1 row) -- lock shard metadata: take some share locks and exclusive locks BEGIN; SELECT lock_shard_metadata(5, ARRAY[999001, 999002, 999002]); lock_shard_metadata --------------------- (1 row) SELECT lock_shard_metadata(7, ARRAY[999001, 999003, 999004]); lock_shard_metadata --------------------- (1 row) SELECT locktype, objid, mode, granted FROM pg_locks WHERE objid IN (999001, 999002, 999003, 999004) ORDER BY objid, mode; locktype | objid | mode | granted ----------+--------+---------------+--------- advisory | 999001 | ExclusiveLock | t advisory | 999001 | ShareLock | t advisory | 999002 | ShareLock | t advisory | 999003 | ExclusiveLock | t advisory | 999004 | ExclusiveLock | t (5 rows) END; -- lock shard metadata: unsupported lock type SELECT lock_shard_metadata(0, ARRAY[990001, 999002]); ERROR: unsupported lockmode 0 -- lock shard metadata: invalid shard ID SELECT lock_shard_metadata(5, ARRAY[0]); lock_shard_metadata --------------------- (1 row) -- lock shard metadata: lock nothing SELECT lock_shard_metadata(5, ARRAY[]::bigint[]); ERROR: no locks specified -- lock shard resources: take some share locks and exclusive locks BEGIN; SELECT lock_shard_resources(5, ARRAY[999001, 999002, 999002]); lock_shard_resources ---------------------- (1 row) SELECT lock_shard_resources(7, ARRAY[999001, 999003, 999004]); lock_shard_resources ---------------------- (1 row) SELECT locktype, objid, mode, granted FROM pg_locks WHERE objid IN (999001, 999002, 999003, 999004) ORDER BY objid, mode; locktype | objid | mode | granted ----------+--------+---------------+--------- advisory | 999001 | ExclusiveLock | t advisory | 999001 | ShareLock | t advisory | 999002 | ShareLock | t advisory | 999003 | ExclusiveLock | t advisory | 999004 | ExclusiveLock | t (5 rows) END; -- lock shard metadata: unsupported lock type SELECT lock_shard_resources(0, ARRAY[990001, 999002]); ERROR: unsupported lockmode 0 -- lock shard metadata: invalid shard ID SELECT lock_shard_resources(5, ARRAY[-1]); lock_shard_resources ---------------------- (1 row) -- lock shard metadata: lock nothing SELECT lock_shard_resources(5, ARRAY[]::bigint[]); ERROR: no locks specified -- drop table DROP TABLE sharded_table; -- VACUUM tests -- create a table with a single shard (for convenience) CREATE TABLE dustbunnies (id integer, name text, age integer); SELECT master_create_distributed_table('dustbunnies', 'id', 'hash'); master_create_distributed_table --------------------------------- (1 row) SELECT master_create_worker_shards('dustbunnies', 1, 2); master_create_worker_shards ----------------------------- (1 row) -- add some data to the distributed table \copy dustbunnies (id, name) from stdin with csv -- following approach adapted from PostgreSQL's stats.sql file -- save relevant stat counter values in refreshable view \c - - - :worker_1_port CREATE MATERIALIZED VIEW prevcounts AS SELECT analyze_count, vacuum_count FROM pg_stat_user_tables WHERE relname='dustbunnies_990002'; -- create function that sleeps until those counters increment create function wait_for_stats() returns void as $$ declare start_time timestamptz := clock_timestamp(); analyze_updated bool; vacuum_updated bool; begin -- we don't want to wait forever; loop will exit after 10 seconds for i in 1 .. 100 loop -- check to see if analyze has been updated SELECT (st.analyze_count >= pc.analyze_count + 1) INTO analyze_updated FROM pg_stat_user_tables AS st, pg_class AS cl, prevcounts AS pc WHERE st.relname='dustbunnies_990002' AND cl.relname='dustbunnies_990002'; -- check to see if vacuum has been updated SELECT (st.vacuum_count >= pc.vacuum_count + 1) INTO vacuum_updated FROM pg_stat_user_tables AS st, pg_class AS cl, prevcounts AS pc WHERE st.relname='dustbunnies_990002' AND cl.relname='dustbunnies_990002'; exit when analyze_updated or vacuum_updated; -- wait a little perform pg_sleep(0.1); -- reset stats snapshot so we can test again perform pg_stat_clear_snapshot(); end loop; -- report time waited in postmaster log (where it won't change test output) raise log 'wait_for_stats delayed % seconds', extract(epoch from clock_timestamp() - start_time); end $$ language plpgsql; -- run VACUUM and ANALYZE against the table on the master \c - - - :master_port VACUUM dustbunnies; ANALYZE dustbunnies; -- verify that the VACUUM and ANALYZE ran \c - - - :worker_1_port SELECT wait_for_stats(); wait_for_stats ---------------- (1 row) REFRESH MATERIALIZED VIEW prevcounts; SELECT pg_stat_get_vacuum_count('dustbunnies_990002'::regclass); pg_stat_get_vacuum_count -------------------------- 1 (1 row) SELECT pg_stat_get_analyze_count('dustbunnies_990002'::regclass); pg_stat_get_analyze_count --------------------------- 1 (1 row) -- get file node to verify VACUUM FULL SELECT relfilenode AS oldnode FROM pg_class WHERE oid='dustbunnies_990002'::regclass \gset -- send a VACUUM FULL and a VACUUM ANALYZE \c - - - :master_port VACUUM (FULL) dustbunnies; VACUUM ANALYZE dustbunnies; -- verify that relfilenode changed \c - - - :worker_1_port SELECT relfilenode != :oldnode AS table_rewritten FROM pg_class WHERE oid='dustbunnies_990002'::regclass; table_rewritten ----------------- t (1 row) -- verify the VACUUM ANALYZE incremented both vacuum and analyze counts SELECT wait_for_stats(); wait_for_stats ---------------- (1 row) SELECT pg_stat_get_vacuum_count('dustbunnies_990002'::regclass); pg_stat_get_vacuum_count -------------------------- 2 (1 row) SELECT pg_stat_get_analyze_count('dustbunnies_990002'::regclass); pg_stat_get_analyze_count --------------------------- 2 (1 row) -- disable auto-VACUUM for next test ALTER TABLE dustbunnies_990002 SET (autovacuum_enabled = false); SELECT relfrozenxid AS frozenxid FROM pg_class WHERE oid='dustbunnies_990002'::regclass \gset -- send a VACUUM FREEZE after adding a new row \c - - - :master_port INSERT INTO dustbunnies VALUES (5, 'peter'); VACUUM (FREEZE) dustbunnies; -- verify that relfrozenxid increased \c - - - :worker_1_port SELECT relfrozenxid::text::integer > :frozenxid AS frozen_performed FROM pg_class WHERE oid='dustbunnies_990002'::regclass; frozen_performed ------------------ t (1 row) -- check there are no nulls in either column SELECT attname, null_frac FROM pg_stats WHERE tablename = 'dustbunnies_990002' ORDER BY attname; attname | null_frac ---------+----------- age | 1 id | 0 name | 0 (3 rows) -- add NULL values, then perform column-specific ANALYZE \c - - - :master_port INSERT INTO dustbunnies VALUES (6, NULL, NULL); ANALYZE dustbunnies (name); -- verify that name's NULL ratio is updated but age's is not \c - - - :worker_1_port SELECT attname, null_frac FROM pg_stats WHERE tablename = 'dustbunnies_990002' ORDER BY attname; attname | null_frac ---------+----------- age | 1 id | 0 name | 0.166667 (3 rows) \c - - - :master_port -- verify warning for unqualified VACUUM VACUUM; WARNING: not propagating VACUUM command to worker nodes HINT: Provide a specific table in order to VACUUM distributed tables. -- and warning when using targeted VACUUM without DDL propagation SET citus.enable_ddl_propagation to false; VACUUM dustbunnies; WARNING: not propagating VACUUM command to worker nodes HINT: Set citus.enable_ddl_propagation to true in order to send targeted VACUUM commands to worker nodes. SET citus.enable_ddl_propagation to DEFAULT; -- TODO: support VERBOSE -- VACUUM VERBOSE dustbunnies; -- VACUUM (FULL, VERBOSE) dustbunnies; -- ANALYZE VERBOSE dustbunnies;