From 945ce21320fc58bb8a0f9c961723f78689d81b6f Mon Sep 17 00:00:00 2001 From: Jason Petersen Date: Mon, 22 Aug 2016 13:24:18 -0600 Subject: [PATCH] Replace verb 'stage' with 'load' in test comments "Staging table" will be the only valid use of 'stage' from now on, we will now say "load" when talking about data ingestion. If creation of shards is its own step, we'll just say "shard creation". --- src/test/regress/expected/multi_hash_pruning.out | 2 +- src/test/regress/expected/multi_hash_pruning_0.out | 2 +- src/test/regress/expected/multi_join_order_tpch_large.out | 2 +- src/test/regress/expected/multi_shard_modify.out | 2 +- src/test/regress/input/multi_agg_distinct.source | 2 +- src/test/regress/input/multi_agg_type_conversion.source | 2 +- src/test/regress/input/multi_append_table_to_shard.source | 6 +++--- src/test/regress/input/multi_large_shardid.source | 4 ++-- src/test/regress/input/multi_master_delete_protocol.source | 2 +- src/test/regress/input/multi_stage_data.source | 2 +- src/test/regress/input/multi_stage_large_records.source | 2 +- src/test/regress/input/multi_stage_more_data.source | 4 ++-- src/test/regress/input/multi_subquery.source | 2 +- src/test/regress/output/multi_agg_distinct.source | 2 +- src/test/regress/output/multi_agg_type_conversion.source | 2 +- src/test/regress/output/multi_append_table_to_shard.source | 6 +++--- src/test/regress/output/multi_large_shardid.source | 4 ++-- src/test/regress/output/multi_master_delete_protocol.source | 2 +- src/test/regress/output/multi_stage_data.source | 2 +- src/test/regress/output/multi_stage_large_records.source | 2 +- src/test/regress/output/multi_stage_more_data.source | 4 ++-- src/test/regress/output/multi_subquery.source | 2 +- src/test/regress/sql/multi_hash_pruning.sql | 2 +- src/test/regress/sql/multi_join_order_tpch_large.sql | 2 +- src/test/regress/sql/multi_shard_modify.sql | 2 +- 25 files changed, 33 insertions(+), 33 deletions(-) diff --git a/src/test/regress/expected/multi_hash_pruning.out b/src/test/regress/expected/multi_hash_pruning.out index a6f5a0688..de473e728 100644 --- a/src/test/regress/expected/multi_hash_pruning.out +++ b/src/test/regress/expected/multi_hash_pruning.out @@ -5,7 +5,7 @@ ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 630000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 630000; -- Create a table partitioned on integer column and update partition type to --- hash. Then stage data to this table and update shard min max values with +-- hash. Then load data into this table and update shard min max values with -- hashed ones. Hash value of 1, 2, 3 and 4 are consecutively -1905060026, -- 1134484726, -28094569 and -1011077333. CREATE TABLE orders_hash_partitioned ( diff --git a/src/test/regress/expected/multi_hash_pruning_0.out b/src/test/regress/expected/multi_hash_pruning_0.out index 35b3ea99b..c23e37ea1 100644 --- a/src/test/regress/expected/multi_hash_pruning_0.out +++ b/src/test/regress/expected/multi_hash_pruning_0.out @@ -5,7 +5,7 @@ ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 630000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 630000; -- Create a table partitioned on integer column and update partition type to --- hash. Then stage data to this table and update shard min max values with +-- hash. Then load data into this table and update shard min max values with -- hashed ones. Hash value of 1, 2, 3 and 4 are consecutively -1905060026, -- 1134484726, -28094569 and -1011077333. CREATE TABLE orders_hash_partitioned ( diff --git a/src/test/regress/expected/multi_join_order_tpch_large.out b/src/test/regress/expected/multi_join_order_tpch_large.out index 6532d9ea1..327687eac 100644 --- a/src/test/regress/expected/multi_join_order_tpch_large.out +++ b/src/test/regress/expected/multi_join_order_tpch_large.out @@ -9,7 +9,7 @@ SET citus.log_multi_join_order TO TRUE; SET client_min_messages TO LOG; -- Change configuration to treat lineitem, orders, customer, and part tables as -- large. The following queries are basically the same as the ones in tpch_small --- except that more data has been staged to customer and part tables. Therefore, +-- except that more data has been loaded into customer and part tables. Therefore, -- we will apply different distributed join strategies for these queries. SET citus.large_table_shard_count TO 2; -- Query #6 from the TPC-H decision support benchmark diff --git a/src/test/regress/expected/multi_shard_modify.out b/src/test/regress/expected/multi_shard_modify.out index fc4866d04..d722c3ad3 100644 --- a/src/test/regress/expected/multi_shard_modify.out +++ b/src/test/regress/expected/multi_shard_modify.out @@ -3,7 +3,7 @@ -- ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 350000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 350000; --- Create a new hash partitioned multi_shard_modify_test table and stage data into it. +-- Create a new hash partitioned multi_shard_modify_test table and load data into it. CREATE TABLE multi_shard_modify_test ( t_key integer not null, t_name varchar(25) not null, diff --git a/src/test/regress/input/multi_agg_distinct.source b/src/test/regress/input/multi_agg_distinct.source index ec6f525ed..a63ecbb05 100644 --- a/src/test/regress/input/multi_agg_distinct.source +++ b/src/test/regress/input/multi_agg_distinct.source @@ -7,7 +7,7 @@ ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 200000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 200000; --- Create a new range partitioned lineitem table and stage data into it +-- Create a new range partitioned lineitem table and load data into it CREATE TABLE lineitem_range ( l_orderkey bigint not null, l_partkey integer not null, diff --git a/src/test/regress/input/multi_agg_type_conversion.source b/src/test/regress/input/multi_agg_type_conversion.source index 8ef9d2b94..71bdc7e15 100644 --- a/src/test/regress/input/multi_agg_type_conversion.source +++ b/src/test/regress/input/multi_agg_type_conversion.source @@ -14,7 +14,7 @@ SELECT sum(l_suppkey) / 2::numeric FROM lineitem; SELECT sum(l_suppkey)::int8 / 2 FROM lineitem; --- Create a new table to test type conversions on different types, and stage +-- Create a new table to test type conversions on different types, and load -- data into this table. Then, apply aggregate functions and divide / multiply -- the results to test type conversions. diff --git a/src/test/regress/input/multi_append_table_to_shard.source b/src/test/regress/input/multi_append_table_to_shard.source index fef9e8488..def84b299 100644 --- a/src/test/regress/input/multi_append_table_to_shard.source +++ b/src/test/regress/input/multi_append_table_to_shard.source @@ -111,7 +111,7 @@ SELECT master_create_empty_shard('multi_append_table_to_shard_date'); SELECT * FROM multi_append_table_to_shard_date; --- Stage an empty table and check that we can query the distributed table +-- Create an empty distributed table and check that we can query it CREATE TABLE multi_append_table_to_shard_stage (LIKE multi_append_table_to_shard_date); SELECT master_append_table_to_shard(shardid, 'multi_append_table_to_shard_stage', 'localhost', 57636) FROM @@ -120,7 +120,7 @@ WHERE 'multi_append_table_to_shard_date'::regclass::oid = logicalrelid; SELECT * FROM multi_append_table_to_shard_date; --- Stage NULL values and check that we can query the table +-- INSERT NULL values and check that we can query the table INSERT INTO multi_append_table_to_shard_stage VALUES (NULL, NULL); SELECT master_append_table_to_shard(shardid, 'multi_append_table_to_shard_stage', 'localhost', 57636) FROM @@ -129,7 +129,7 @@ WHERE 'multi_append_table_to_shard_date'::regclass::oid = logicalrelid; SELECT * FROM multi_append_table_to_shard_date; --- Stage regular values and check that we can query the table +-- INSERT regular values and check that we can query the table INSERT INTO multi_append_table_to_shard_stage VALUES ('2016-01-01', 3); SELECT master_append_table_to_shard(shardid, 'multi_append_table_to_shard_stage', 'localhost', 57636) FROM diff --git a/src/test/regress/input/multi_large_shardid.source b/src/test/regress/input/multi_large_shardid.source index 53767e90c..2ce50e190 100644 --- a/src/test/regress/input/multi_large_shardid.source +++ b/src/test/regress/input/multi_large_shardid.source @@ -2,7 +2,7 @@ -- MULTI_LARGE_SHARDID -- --- Stage data to distributed tables, and run TPC-H query #1 and #6. This test +-- Load data into distributed tables, and run TPC-H query #1 and #6. This test -- differs from previous tests in that it modifies the *internal* shardId -- generator, forcing the distributed database to use 64-bit shard identifiers. @@ -11,7 +11,7 @@ ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 100200300400500; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 9000000; --- Stage additional data to start using large shard identifiers. +-- Load additional data to start using large shard identifiers. \copy lineitem FROM '@abs_srcdir@/data/lineitem.1.data' with delimiter '|' \copy lineitem FROM '@abs_srcdir@/data/lineitem.2.data' with delimiter '|' diff --git a/src/test/regress/input/multi_master_delete_protocol.source b/src/test/regress/input/multi_master_delete_protocol.source index 4160102e6..378985b63 100644 --- a/src/test/regress/input/multi_master_delete_protocol.source +++ b/src/test/regress/input/multi_master_delete_protocol.source @@ -7,7 +7,7 @@ ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 320000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 320000; --- Create a new range partitioned customer_delete_protocol table and stage data into it. +-- Create a new range partitioned customer_delete_protocol table and load data into it. CREATE TABLE customer_delete_protocol ( c_custkey integer not null, c_name varchar(25) not null, diff --git a/src/test/regress/input/multi_stage_data.source b/src/test/regress/input/multi_stage_data.source index 850b4edef..6daf16d3e 100644 --- a/src/test/regress/input/multi_stage_data.source +++ b/src/test/regress/input/multi_stage_data.source @@ -1,7 +1,7 @@ -- -- MULTI_STAGE_DATA -- --- Tests for staging data in a distributed cluster. Please note that the number +-- Tests for loading data in a distributed cluster. Please note that the number -- of shards uploaded depends on two config values: citus.shard_replication_factor and -- citus.shard_max_size. These values are set in pg_regress_multi.pl. Shard placement -- policy is left to the default value (round-robin) to test the common install case. diff --git a/src/test/regress/input/multi_stage_large_records.source b/src/test/regress/input/multi_stage_large_records.source index 18c73b55e..d94396ce8 100644 --- a/src/test/regress/input/multi_stage_large_records.source +++ b/src/test/regress/input/multi_stage_large_records.source @@ -1,7 +1,7 @@ -- -- MULTI_STAGE_LARGE_RECORDS -- --- Tests for staging data with large records (i.e. greater than the read buffer +-- Tests for loading data with large records (i.e. greater than the read buffer -- size, which is 32kB) in a distributed cluster. These tests make sure that we -- are creating shards of correct size even when records are large. diff --git a/src/test/regress/input/multi_stage_more_data.source b/src/test/regress/input/multi_stage_more_data.source index 651ce2625..655201ac9 100644 --- a/src/test/regress/input/multi_stage_more_data.source +++ b/src/test/regress/input/multi_stage_more_data.source @@ -7,8 +7,8 @@ ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 280000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 280000; --- We stage more data to customer and part tables to test distributed joins. The --- staging causes the planner to consider customer and part tables as large, and +-- We load more data to customer and part tables to test distributed joins. The +-- loading causes the planner to consider customer and part tables as large, and -- evaluate plans where some of the underlying tables need to be repartitioned. \copy customer FROM '@abs_srcdir@/data/customer.2.data' with delimiter '|' diff --git a/src/test/regress/input/multi_subquery.source b/src/test/regress/input/multi_subquery.source index 5976e2dff..e3f684b34 100644 --- a/src/test/regress/input/multi_subquery.source +++ b/src/test/regress/input/multi_subquery.source @@ -77,7 +77,7 @@ FROM GROUP BY l_orderkey) AS unit_prices; --- Stage data to tables. +-- Load data into tables. SELECT master_create_empty_shard('lineitem_subquery') AS new_shard_id \gset diff --git a/src/test/regress/output/multi_agg_distinct.source b/src/test/regress/output/multi_agg_distinct.source index cb75d8f23..41bc722ee 100644 --- a/src/test/regress/output/multi_agg_distinct.source +++ b/src/test/regress/output/multi_agg_distinct.source @@ -3,7 +3,7 @@ -- ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 200000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 200000; --- Create a new range partitioned lineitem table and stage data into it +-- Create a new range partitioned lineitem table and load data into it CREATE TABLE lineitem_range ( l_orderkey bigint not null, l_partkey integer not null, diff --git a/src/test/regress/output/multi_agg_type_conversion.source b/src/test/regress/output/multi_agg_type_conversion.source index 1ba41083b..82a30be1a 100644 --- a/src/test/regress/output/multi_agg_type_conversion.source +++ b/src/test/regress/output/multi_agg_type_conversion.source @@ -28,7 +28,7 @@ SELECT sum(l_suppkey)::int8 / 2 FROM lineitem; 30308988 (1 row) --- Create a new table to test type conversions on different types, and stage +-- Create a new table to test type conversions on different types, and load -- data into this table. Then, apply aggregate functions and divide / multiply -- the results to test type conversions. CREATE TABLE aggregate_type ( diff --git a/src/test/regress/output/multi_append_table_to_shard.source b/src/test/regress/output/multi_append_table_to_shard.source index c04a9563d..8bf9f69da 100644 --- a/src/test/regress/output/multi_append_table_to_shard.source +++ b/src/test/regress/output/multi_append_table_to_shard.source @@ -175,7 +175,7 @@ SELECT * FROM multi_append_table_to_shard_date; ------------+------- (0 rows) --- Stage an empty table and check that we can query the distributed table +-- Create an empty distributed table and check that we can query it CREATE TABLE multi_append_table_to_shard_stage (LIKE multi_append_table_to_shard_date); SELECT master_append_table_to_shard(shardid, 'multi_append_table_to_shard_stage', 'localhost', 57636) FROM @@ -191,7 +191,7 @@ SELECT * FROM multi_append_table_to_shard_date; ------------+------- (0 rows) --- Stage NULL values and check that we can query the table +-- INSERT NULL values and check that we can query the table INSERT INTO multi_append_table_to_shard_stage VALUES (NULL, NULL); SELECT master_append_table_to_shard(shardid, 'multi_append_table_to_shard_stage', 'localhost', 57636) FROM @@ -208,7 +208,7 @@ SELECT * FROM multi_append_table_to_shard_date; | (1 row) --- Stage regular values and check that we can query the table +-- INSERT regular values and check that we can query the table INSERT INTO multi_append_table_to_shard_stage VALUES ('2016-01-01', 3); SELECT master_append_table_to_shard(shardid, 'multi_append_table_to_shard_stage', 'localhost', 57636) FROM diff --git a/src/test/regress/output/multi_large_shardid.source b/src/test/regress/output/multi_large_shardid.source index ac291ca78..d223875b0 100644 --- a/src/test/regress/output/multi_large_shardid.source +++ b/src/test/regress/output/multi_large_shardid.source @@ -1,12 +1,12 @@ -- -- MULTI_LARGE_SHARDID -- --- Stage data to distributed tables, and run TPC-H query #1 and #6. This test +-- Load data into distributed tables, and run TPC-H query #1 and #6. This test -- differs from previous tests in that it modifies the *internal* shardId -- generator, forcing the distributed database to use 64-bit shard identifiers. ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 100200300400500; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 9000000; --- Stage additional data to start using large shard identifiers. +-- Load additional data to start using large shard identifiers. \copy lineitem FROM '@abs_srcdir@/data/lineitem.1.data' with delimiter '|' \copy lineitem FROM '@abs_srcdir@/data/lineitem.2.data' with delimiter '|' -- Query #1 from the TPC-H decision support benchmark. diff --git a/src/test/regress/output/multi_master_delete_protocol.source b/src/test/regress/output/multi_master_delete_protocol.source index a1499b273..0e35f9b4d 100644 --- a/src/test/regress/output/multi_master_delete_protocol.source +++ b/src/test/regress/output/multi_master_delete_protocol.source @@ -3,7 +3,7 @@ -- ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 320000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 320000; --- Create a new range partitioned customer_delete_protocol table and stage data into it. +-- Create a new range partitioned customer_delete_protocol table and load data into it. CREATE TABLE customer_delete_protocol ( c_custkey integer not null, c_name varchar(25) not null, diff --git a/src/test/regress/output/multi_stage_data.source b/src/test/regress/output/multi_stage_data.source index 0cc434c59..b27ff1e3b 100644 --- a/src/test/regress/output/multi_stage_data.source +++ b/src/test/regress/output/multi_stage_data.source @@ -1,7 +1,7 @@ -- -- MULTI_STAGE_DATA -- --- Tests for staging data in a distributed cluster. Please note that the number +-- Tests for loading data in a distributed cluster. Please note that the number -- of shards uploaded depends on two config values: citus.shard_replication_factor and -- citus.shard_max_size. These values are set in pg_regress_multi.pl. Shard placement -- policy is left to the default value (round-robin) to test the common install case. diff --git a/src/test/regress/output/multi_stage_large_records.source b/src/test/regress/output/multi_stage_large_records.source index 918a30401..70a53cf26 100644 --- a/src/test/regress/output/multi_stage_large_records.source +++ b/src/test/regress/output/multi_stage_large_records.source @@ -1,7 +1,7 @@ -- -- MULTI_STAGE_LARGE_RECORDS -- --- Tests for staging data with large records (i.e. greater than the read buffer +-- Tests for loading data with large records (i.e. greater than the read buffer -- size, which is 32kB) in a distributed cluster. These tests make sure that we -- are creating shards of correct size even when records are large. ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 300000; diff --git a/src/test/regress/output/multi_stage_more_data.source b/src/test/regress/output/multi_stage_more_data.source index 20dd3e89a..dec7d9c26 100644 --- a/src/test/regress/output/multi_stage_more_data.source +++ b/src/test/regress/output/multi_stage_more_data.source @@ -3,8 +3,8 @@ -- ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 280000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 280000; --- We stage more data to customer and part tables to test distributed joins. The --- staging causes the planner to consider customer and part tables as large, and +-- We load more data to customer and part tables to test distributed joins. The +-- loading causes the planner to consider customer and part tables as large, and -- evaluate plans where some of the underlying tables need to be repartitioned. \copy customer FROM '@abs_srcdir@/data/customer.2.data' with delimiter '|' \copy customer FROM '@abs_srcdir@/data/customer.3.data' with delimiter '|' diff --git a/src/test/regress/output/multi_subquery.source b/src/test/regress/output/multi_subquery.source index 992295706..e0340f3f6 100644 --- a/src/test/regress/output/multi_subquery.source +++ b/src/test/regress/output/multi_subquery.source @@ -82,7 +82,7 @@ FROM (1 row) --- Stage data to tables. +-- Load data into tables. SELECT master_create_empty_shard('lineitem_subquery') AS new_shard_id \gset UPDATE pg_dist_shard SET shardminvalue = 1, shardmaxvalue = 5986 diff --git a/src/test/regress/sql/multi_hash_pruning.sql b/src/test/regress/sql/multi_hash_pruning.sql index d6210d8f8..0088264a4 100644 --- a/src/test/regress/sql/multi_hash_pruning.sql +++ b/src/test/regress/sql/multi_hash_pruning.sql @@ -10,7 +10,7 @@ ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 630000; -- Create a table partitioned on integer column and update partition type to --- hash. Then stage data to this table and update shard min max values with +-- hash. Then load data into this table and update shard min max values with -- hashed ones. Hash value of 1, 2, 3 and 4 are consecutively -1905060026, -- 1134484726, -28094569 and -1011077333. diff --git a/src/test/regress/sql/multi_join_order_tpch_large.sql b/src/test/regress/sql/multi_join_order_tpch_large.sql index 89ab65411..542b33385 100644 --- a/src/test/regress/sql/multi_join_order_tpch_large.sql +++ b/src/test/regress/sql/multi_join_order_tpch_large.sql @@ -15,7 +15,7 @@ SET client_min_messages TO LOG; -- Change configuration to treat lineitem, orders, customer, and part tables as -- large. The following queries are basically the same as the ones in tpch_small --- except that more data has been staged to customer and part tables. Therefore, +-- except that more data has been loaded into customer and part tables. Therefore, -- we will apply different distributed join strategies for these queries. SET citus.large_table_shard_count TO 2; diff --git a/src/test/regress/sql/multi_shard_modify.sql b/src/test/regress/sql/multi_shard_modify.sql index f8bd8f88c..9b1fabf26 100644 --- a/src/test/regress/sql/multi_shard_modify.sql +++ b/src/test/regress/sql/multi_shard_modify.sql @@ -7,7 +7,7 @@ ALTER SEQUENCE pg_catalog.pg_dist_shardid_seq RESTART 350000; ALTER SEQUENCE pg_catalog.pg_dist_jobid_seq RESTART 350000; --- Create a new hash partitioned multi_shard_modify_test table and stage data into it. +-- Create a new hash partitioned multi_shard_modify_test table and load data into it. CREATE TABLE multi_shard_modify_test ( t_key integer not null, t_name varchar(25) not null,