Fix typos:

VAR_SET_VALUE_KIND -> VAR_SET_VALUE kind
beginnig -> beginning
plannig -> planning
the the -> the
er then -> er than
pull/3227/head
Philip Dubé 2019-11-25 23:06:33 +00:00
parent 4b0ac4b0dd
commit 261a9de42d
24 changed files with 30 additions and 30 deletions

View File

@ -231,7 +231,7 @@ ReplicateAllDependenciesToNode(const char *nodeName, int nodePort)
/* /*
* When dependency lists are getting longer we see a delay in the creation time on the * When dependency lists are getting longer we see a delay in the creation time on the
* workers. We would like to inform the user. Currently we warn for lists greater then * workers. We would like to inform the user. Currently we warn for lists greater than
* 100 items, where 100 is an arbitrarily chosen number. If we find it too high or too * 100 items, where 100 is an arbitrarily chosen number. If we find it too high or too
* low we can adjust this based on experience. * low we can adjust this based on experience.
*/ */

View File

@ -522,7 +522,7 @@ PlanAlterExtensionUpdateStmt(AlterExtensionStmt *alterExtensionStmt, const
* EnsureSequentialModeForExtensionDDL makes sure that the current transaction is already in * EnsureSequentialModeForExtensionDDL makes sure that the current transaction is already in
* sequential mode, or can still safely be put in sequential mode, it errors if that is * sequential mode, or can still safely be put in sequential mode, it errors if that is
* not possible. The error contains information for the user to retry the transaction with * not possible. The error contains information for the user to retry the transaction with
* sequential mode set from the beginnig. * sequential mode set from the beginning.
* *
* As extensions are node scoped objects there exists only 1 instance of the * As extensions are node scoped objects there exists only 1 instance of the
* extension used by potentially multiple shards. To make sure all shards in * extension used by potentially multiple shards. To make sure all shards in

View File

@ -989,7 +989,7 @@ GetAggregateDDLCommand(const RegProcedure funcOid, bool useCreateOrReplace)
* EnsureSequentialModeForFunctionDDL makes sure that the current transaction is already in * EnsureSequentialModeForFunctionDDL makes sure that the current transaction is already in
* sequential mode, or can still safely be put in sequential mode, it errors if that is * sequential mode, or can still safely be put in sequential mode, it errors if that is
* not possible. The error contains information for the user to retry the transaction with * not possible. The error contains information for the user to retry the transaction with
* sequential mode set from the beginnig. * sequential mode set from the beginning.
* *
* As functions are node scoped objects there exists only 1 instance of the function used by * As functions are node scoped objects there exists only 1 instance of the function used by
* potentially multiple shards. To make sure all shards in the transaction can interact * potentially multiple shards. To make sure all shards in the transaction can interact

View File

@ -368,7 +368,7 @@ PlanAlterEnumStmt(AlterEnumStmt *stmt, const char *queryString)
/* /*
* ProcessAlterEnumStmt is called after the AlterEnumStmt has been applied locally. * ProcessAlterEnumStmt is called after the AlterEnumStmt has been applied locally.
* *
* This function is used for ALTER ENUM ... ADD VALUE for postgres versions lower then 12 * This function is used for ALTER ENUM ... ADD VALUE for postgres versions lower than 12
* to distribute the call. Before pg12 these statements could not be called in a * to distribute the call. Before pg12 these statements could not be called in a
* transaction. If we would plan the distirbution of these statements the same as we do * transaction. If we would plan the distirbution of these statements the same as we do
* with the other statements they would get executed in a transaction to perform 2PC, that * with the other statements they would get executed in a transaction to perform 2PC, that

View File

@ -695,7 +695,7 @@ multi_ProcessUtility(PlannedStmt *pstmt,
* CommandCounterIncrement twice, as the call is a no-op if the command id is not * CommandCounterIncrement twice, as the call is a no-op if the command id is not
* used yet. * used yet.
* *
* Once versions older then above are not deemed important anymore this patch can * Once versions older than above are not deemed important anymore this patch can
* be remove from citus. * be remove from citus.
*/ */
CommandCounterIncrement(); CommandCounterIncrement();

View File

@ -315,7 +315,7 @@ AppendDefElemSet(StringInfo buf, DefElem *def)
/* /*
* AppendVarSetValue deparses a VariableSetStmt with VAR_SET_VALUE_KIND * AppendVarSetValue deparses a VariableSetStmt with VAR_SET_VALUE kind.
* It takes from flatten_set_variable_args in postgres's utils/misc/guc.c, * It takes from flatten_set_variable_args in postgres's utils/misc/guc.c,
* however flatten_set_variable_args does not apply correct quoting. * however flatten_set_variable_args does not apply correct quoting.
*/ */

View File

@ -2771,7 +2771,7 @@ UpdateConnectionWaitFlags(WorkerSession *session, int waitFlags)
/* /*
* CheckConnectionReady returns true if the the connection is ready to * CheckConnectionReady returns true if the connection is ready to
* read or write, or false if it still has bytes to send/receive. * read or write, or false if it still has bytes to send/receive.
*/ */
static bool static bool

View File

@ -1620,7 +1620,7 @@ CheckAvailableVersion(int elevel)
/* /*
* CheckInstalledVersion compares CITUS_EXTENSIONVERSION and the the * CheckInstalledVersion compares CITUS_EXTENSIONVERSION and the
* extension's current version from the pg_extemsion catalog table. If they * extension's current version from the pg_extemsion catalog table. If they
* are not compatible, this function logs an error with the specified elevel, * are not compatible, this function logs an error with the specified elevel,
* otherwise it returns true. * otherwise it returns true.

View File

@ -1707,7 +1707,7 @@ CreateAndPushPlannerRestrictionContext(void)
/* /*
* CurrentRestrictionContext returns the the most recently added * CurrentRestrictionContext returns the most recently added
* PlannerRestrictionContext from the plannerRestrictionContextList list. * PlannerRestrictionContext from the plannerRestrictionContextList list.
*/ */
static PlannerRestrictionContext * static PlannerRestrictionContext *

View File

@ -366,7 +366,7 @@ ExplainTaskList(List *taskList, ExplainState *es)
/* /*
* RemoteExplain fetches the the remote EXPLAIN output for a single * RemoteExplain fetches the remote EXPLAIN output for a single
* task. It tries each shard placement until one succeeds or all * task. It tries each shard placement until one succeeds or all
* failed. * failed.
*/ */

View File

@ -2322,7 +2322,7 @@ ProcessHavingClauseForWorkerQuery(Node *originalHavingQual,
/* /*
* PrcoessDistinctClauseForWorkerQuery gets the inputs and modifies the outputs * PrcoessDistinctClauseForWorkerQuery gets the inputs and modifies the outputs
* such that worker query's DISTINCT and DISTINCT ON clauses are set accordingly. * such that worker query's DISTINCT and DISTINCT ON clauses are set accordingly.
* Note the the function may or may not decide to pushdown the DISTINCT and DISTINCT * Note the function may or may not decide to pushdown the DISTINCT and DISTINCT
* on clauses based on the inputs. * on clauses based on the inputs.
* *
* See the detailed comments in the function for the rules of pushing down DISTINCT * See the detailed comments in the function for the rules of pushing down DISTINCT

View File

@ -90,7 +90,7 @@ typedef struct InsertValues
/* /*
* A ModifyRoute encapsulates the the information needed to route modifications * A ModifyRoute encapsulates the information needed to route modifications
* to the appropriate shard. For a single-shard modification, only one route * to the appropriate shard. For a single-shard modification, only one route
* is needed, but in the case of e.g. a multi-row INSERT, lists of these values * is needed, but in the case of e.g. a multi-row INSERT, lists of these values
* will help divide the rows by their destination shards, permitting later * will help divide the rows by their destination shards, permitting later
@ -2429,7 +2429,7 @@ WorkersContainingAllShards(List *prunedShardIntervalsList)
/* /*
* Bail out if placement list becomes empty. This means there is no worker * Bail out if placement list becomes empty. This means there is no worker
* containing all shards referecend by the query, hence we can not forward * containing all shards referenced by the query, hence we can not forward
* this query directly to any worker. * this query directly to any worker.
*/ */
if (currentPlacementList == NIL) if (currentPlacementList == NIL)

View File

@ -462,7 +462,7 @@ PrunableExpressions(Node *node, ClauseWalkerContext *context)
* be pending. * be pending.
* *
* We copy the partial PruningInstance, and continue adding information by * We copy the partial PruningInstance, and continue adding information by
* calling PrunableExpressionsWalker() on the copy, continuing at the the * calling PrunableExpressionsWalker() on the copy, continuing at the
* node stored in PendingPruningInstance->continueAt. * node stored in PendingPruningInstance->continueAt.
*/ */
while (context->pendingInstances != NIL) while (context->pendingInstances != NIL)

View File

@ -255,7 +255,7 @@ StartRemoteTransactionCommit(MultiConnection *connection)
/* /*
* For a moment there I thought we were in trouble. * For a moment there I thought we were in trouble.
* *
* Failing in this state means that we don't know whether the the * Failing in this state means that we don't know whether the
* commit has succeeded. * commit has succeeded.
*/ */
HandleRemoteTransactionConnectionError(connection, raiseErrors); HandleRemoteTransactionConnectionError(connection, raiseErrors);
@ -288,7 +288,7 @@ FinishRemoteTransactionCommit(MultiConnection *connection)
/* /*
* Failing in this state means that we will often not know whether * Failing in this state means that we will often not know whether
* the the commit has succeeded (particularly in case of network * the commit has succeeded (particularly in case of network
* troubles). * troubles).
* *
* XXX: It might be worthwhile to discern cases where we got a * XXX: It might be worthwhile to discern cases where we got a

View File

@ -65,7 +65,7 @@ typedef enum AdvisoryLocktagClass
ADV_LOCKTAG_CLASS_CITUS_JOB) ADV_LOCKTAG_CLASS_CITUS_JOB)
/* reuse advisory lock, but with different, unused field 4 (7) /* reuse advisory lock, but with different, unused field 4 (7)
* Also it has the the database hardcoded to MyDatabaseId, to ensure the locks * Also it has the database hardcoded to MyDatabaseId, to ensure the locks
* are local to each database */ * are local to each database */
#define SET_LOCKTAG_REBALANCE_COLOCATION(tag, colocationOrTableId) \ #define SET_LOCKTAG_REBALANCE_COLOCATION(tag, colocationOrTableId) \
SET_LOCKTAG_ADVISORY(tag, \ SET_LOCKTAG_ADVISORY(tag, \

View File

@ -1986,7 +1986,7 @@ FROM
DEBUG: distributed statement: INSERT INTO public.table_with_defaults_13300017 AS citus_table_alias (store_id, first_name, default_1, last_name, default_2) SELECT store_id, 'Andres'::text AS first_name, 10, 'Freund'::text AS last_name, 20 FROM public.table_with_defaults_13300017 table_with_defaults WHERE ((worker_hash(store_id) OPERATOR(pg_catalog.>=) '-2147483648'::integer) AND (worker_hash(store_id) OPERATOR(pg_catalog.<=) '-1'::integer)) DEBUG: distributed statement: INSERT INTO public.table_with_defaults_13300017 AS citus_table_alias (store_id, first_name, default_1, last_name, default_2) SELECT store_id, 'Andres'::text AS first_name, 10, 'Freund'::text AS last_name, 20 FROM public.table_with_defaults_13300017 table_with_defaults WHERE ((worker_hash(store_id) OPERATOR(pg_catalog.>=) '-2147483648'::integer) AND (worker_hash(store_id) OPERATOR(pg_catalog.<=) '-1'::integer))
DEBUG: distributed statement: INSERT INTO public.table_with_defaults_13300018 AS citus_table_alias (store_id, first_name, default_1, last_name, default_2) SELECT store_id, 'Andres'::text AS first_name, 10, 'Freund'::text AS last_name, 20 FROM public.table_with_defaults_13300018 table_with_defaults WHERE ((worker_hash(store_id) OPERATOR(pg_catalog.>=) 0) AND (worker_hash(store_id) OPERATOR(pg_catalog.<=) 2147483647)) DEBUG: distributed statement: INSERT INTO public.table_with_defaults_13300018 AS citus_table_alias (store_id, first_name, default_1, last_name, default_2) SELECT store_id, 'Andres'::text AS first_name, 10, 'Freund'::text AS last_name, 20 FROM public.table_with_defaults_13300018 table_with_defaults WHERE ((worker_hash(store_id) OPERATOR(pg_catalog.>=) 0) AND (worker_hash(store_id) OPERATOR(pg_catalog.<=) 2147483647))
DEBUG: Plan is router executable DEBUG: Plan is router executable
-- some of the the ultimate queries where we have constants, -- some of the ultimate queries where we have constants,
-- defaults and group by entry is not on the target entry -- defaults and group by entry is not on the target entry
INSERT INTO table_with_defaults (default_2, store_id, first_name) INSERT INTO table_with_defaults (default_2, store_id, first_name)
SELECT SELECT

View File

@ -25,7 +25,7 @@ SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 903
-- We use the l_linenumber field for the following aggregations. We need to use -- We use the l_linenumber field for the following aggregations. We need to use
-- an integer type, as aggregations on numerics or big integers return numerics -- an integer type, as aggregations on numerics or big integers return numerics
-- of unknown length. When the numerics are read into our temporary table, they -- of unknown length. When the numerics are read into our temporary table, they
-- trigger the the creation of toasted tables and indexes. This in turn prints -- trigger the creation of toasted tables and indexes. This in turn prints
-- non-deterministic debug messages. To avoid this chain, we use l_linenumber. -- non-deterministic debug messages. To avoid this chain, we use l_linenumber.
SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030; SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030;
sum | avg sum | avg

View File

@ -765,7 +765,7 @@ DEBUG: Plan 80 query after replacing subqueries and CTEs: SELECT a.user_id, foo
t t
(1 row) (1 row)
-- this time the the plan is optimial, we are -- this time the plan is optimial, we are
-- able to keep the UNION query given that foo -- able to keep the UNION query given that foo
-- is the anchor -- is the anchor
SELECT true AS valid FROM explain_json_2($$ SELECT true AS valid FROM explain_json_2($$
@ -943,7 +943,7 @@ DEBUG: skipping recursive planning for the subquery since it contains reference
DEBUG: skipping recursive planning for the subquery since it contains references to outer queries DEBUG: skipping recursive planning for the subquery since it contains references to outer queries
DEBUG: skipping recursive planning for the subquery since it contains references to outer queries DEBUG: skipping recursive planning for the subquery since it contains references to outer queries
ERROR: complex joins are only supported when all distributed tables are joined on their distribution columns with equal operator ERROR: complex joins are only supported when all distributed tables are joined on their distribution columns with equal operator
-- similar to the above, make sure that we skip recursive plannig when -- similar to the above, make sure that we skip recursive planning when
-- the subquery doesn't have any tables -- the subquery doesn't have any tables
SELECT true AS valid FROM explain_json_2($$ SELECT true AS valid FROM explain_json_2($$
SELECT * SELECT *
@ -962,7 +962,7 @@ DEBUG: skipping recursive planning for the subquery since it contains reference
DEBUG: skipping recursive planning for the subquery since it contains references to outer queries DEBUG: skipping recursive planning for the subquery since it contains references to outer queries
DEBUG: skipping recursive planning for the subquery since it contains references to outer queries DEBUG: skipping recursive planning for the subquery since it contains references to outer queries
ERROR: complex joins are only supported when all distributed tables are joined on their distribution columns with equal operator ERROR: complex joins are only supported when all distributed tables are joined on their distribution columns with equal operator
-- similar to the above, make sure that we skip recursive plannig when -- similar to the above, make sure that we skip recursive planning when
-- the subquery contains only intermediate results -- the subquery contains only intermediate results
SELECT * SELECT *
FROM FROM

View File

@ -24,7 +24,7 @@ SET search_path TO public;
-- These copies were intended to test copying data to single sharded table from -- These copies were intended to test copying data to single sharded table from
-- worker nodes, yet in order to remove broadcast logic related codes we change -- worker nodes, yet in order to remove broadcast logic related codes we change
-- the the table to reference table and copy data from master. Should be updated -- the table to reference table and copy data from master. Should be updated
-- when worker nodes gain capability to run dml commands on reference tables. -- when worker nodes gain capability to run dml commands on reference tables.
\c - - - :master_port \c - - - :master_port
SET search_path TO public; SET search_path TO public;

View File

@ -18,7 +18,7 @@ SET search_path TO public;
\COPY orders_mx FROM '@abs_srcdir@/data/orders.2.data' with delimiter '|' \COPY orders_mx FROM '@abs_srcdir@/data/orders.2.data' with delimiter '|'
-- These copies were intended to test copying data to single sharded table from -- These copies were intended to test copying data to single sharded table from
-- worker nodes, yet in order to remove broadcast logic related codes we change -- worker nodes, yet in order to remove broadcast logic related codes we change
-- the the table to reference table and copy data from master. Should be updated -- the table to reference table and copy data from master. Should be updated
-- when worker nodes gain capability to run dml commands on reference tables. -- when worker nodes gain capability to run dml commands on reference tables.
\c - - - :master_port \c - - - :master_port
SET search_path TO public; SET search_path TO public;

View File

@ -400,7 +400,7 @@ permutation "s1-begin" "s2-begin" "s1-insert-dist-10" "s2-insert-local-10" "s2-i
# daedlock with reference tables only # daedlock with reference tables only
permutation "s1-begin" "s2-begin" "s2-insert-ref-10" "s1-insert-ref-11" "s2-insert-ref-11" "s1-insert-ref-10" "deadlock-checker-call" "s1-commit" "s2-commit" permutation "s1-begin" "s2-begin" "s2-insert-ref-10" "s1-insert-ref-11" "s2-insert-ref-11" "s1-insert-ref-10" "deadlock-checker-call" "s1-commit" "s2-commit"
# deadlock with referecen + distributed tables # deadlock with reference + distributed tables
permutation "s1-begin" "s2-begin" "s2-insert-ref-10" "s1-update-1" "deadlock-checker-call" "s2-update-1" "s1-insert-ref-10" "deadlock-checker-call" "s1-commit" "s2-commit" permutation "s1-begin" "s2-begin" "s2-insert-ref-10" "s1-update-1" "deadlock-checker-call" "s2-update-1" "s1-insert-ref-10" "deadlock-checker-call" "s1-commit" "s2-commit"
# slightly more complex case, loop with three nodes # slightly more complex case, loop with three nodes

View File

@ -1598,7 +1598,7 @@ SELECT
FROM FROM
table_with_defaults; table_with_defaults;
-- some of the the ultimate queries where we have constants, -- some of the ultimate queries where we have constants,
-- defaults and group by entry is not on the target entry -- defaults and group by entry is not on the target entry
INSERT INTO table_with_defaults (default_2, store_id, first_name) INSERT INTO table_with_defaults (default_2, store_id, first_name)
SELECT SELECT

View File

@ -14,7 +14,7 @@ SELECT l_orderkey, l_linenumber, l_shipdate FROM lineitem WHERE l_orderkey = 903
-- We use the l_linenumber field for the following aggregations. We need to use -- We use the l_linenumber field for the following aggregations. We need to use
-- an integer type, as aggregations on numerics or big integers return numerics -- an integer type, as aggregations on numerics or big integers return numerics
-- of unknown length. When the numerics are read into our temporary table, they -- of unknown length. When the numerics are read into our temporary table, they
-- trigger the the creation of toasted tables and indexes. This in turn prints -- trigger the creation of toasted tables and indexes. This in turn prints
-- non-deterministic debug messages. To avoid this chain, we use l_linenumber. -- non-deterministic debug messages. To avoid this chain, we use l_linenumber.
SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030; SELECT sum(l_linenumber), avg(l_linenumber) FROM lineitem WHERE l_orderkey > 9030;

View File

@ -578,7 +578,7 @@ SELECT true AS valid FROM explain_json_2($$
); );
$$); $$);
-- this time the the plan is optimial, we are -- this time the plan is optimial, we are
-- able to keep the UNION query given that foo -- able to keep the UNION query given that foo
-- is the anchor -- is the anchor
SELECT true AS valid FROM explain_json_2($$ SELECT true AS valid FROM explain_json_2($$
@ -707,7 +707,7 @@ JOIN LATERAL
WHERE user_id = users_table.user_id) AS bar WHERE user_id = users_table.user_id) AS bar
LEFT JOIN users_table u2 ON u2.user_id = bar.value_2) AS foo ON TRUE; LEFT JOIN users_table u2 ON u2.user_id = bar.value_2) AS foo ON TRUE;
-- similar to the above, make sure that we skip recursive plannig when -- similar to the above, make sure that we skip recursive planning when
-- the subquery doesn't have any tables -- the subquery doesn't have any tables
SELECT true AS valid FROM explain_json_2($$ SELECT true AS valid FROM explain_json_2($$
SELECT * SELECT *
@ -722,7 +722,7 @@ JOIN LATERAL
LEFT JOIN users_table u2 ON u2.user_id = bar.value_2) AS foo ON TRUE LEFT JOIN users_table u2 ON u2.user_id = bar.value_2) AS foo ON TRUE
$$); $$);
-- similar to the above, make sure that we skip recursive plannig when -- similar to the above, make sure that we skip recursive planning when
-- the subquery contains only intermediate results -- the subquery contains only intermediate results
SELECT * SELECT *
FROM FROM