Merge pull request #4779 from citusdata/typos

Fix various typos due to zealous repetition
pull/4345/head
Philip Dubé 2021-03-05 13:16:12 +00:00 committed by GitHub
commit ce296ac62e
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
13 changed files with 35 additions and 35 deletions

View File

@ -272,7 +272,7 @@ GetConnParams(ConnectionHashKey *key, char ***keywords, char ***values,
/* /*
* Declare local params for readability; * Declare local params for readability;
* *
* assignment is done directly to not loose the pointers if any of the later * assignment is done directly to not lose the pointers if any of the later
* allocations cause an error. FreeConnParamsHashEntryFields knows about the * allocations cause an error. FreeConnParamsHashEntryFields knows about the
* possibility of half initialized keywords or values and correctly reclaims them when * possibility of half initialized keywords or values and correctly reclaims them when
* the cache is reused. * the cache is reused.

View File

@ -370,7 +370,7 @@ StartNodeUserDatabaseConnection(uint32 flags, const char *hostname, int32 port,
* remote node as it might not have any space in * remote node as it might not have any space in
* max_connections for this connection establishment. * max_connections for this connection establishment.
* *
* Still, we keep track of the connnection counter. * Still, we keep track of the connection counter.
*/ */
IncrementSharedConnectionCounter(hostname, port); IncrementSharedConnectionCounter(hostname, port);
} }

View File

@ -219,7 +219,7 @@ ClearResultsIfReady(MultiConnection *connection)
if (!(resultStatus == PGRES_SINGLE_TUPLE || resultStatus == PGRES_TUPLES_OK || if (!(resultStatus == PGRES_SINGLE_TUPLE || resultStatus == PGRES_TUPLES_OK ||
resultStatus == PGRES_COMMAND_OK)) resultStatus == PGRES_COMMAND_OK))
{ {
/* an error occcurred just when we were aborting */ /* an error occurred just when we were aborting */
return false; return false;
} }

View File

@ -15,7 +15,7 @@
* executes "unassigned" tasks from a queue. * executes "unassigned" tasks from a queue.
* - WorkerSession: * - WorkerSession:
* Connection to a worker that is used to execute "assigned" tasks * Connection to a worker that is used to execute "assigned" tasks
* from a queue and may execute unasssigned tasks from the WorkerPool. * from a queue and may execute unassigned tasks from the WorkerPool.
* - ShardCommandExecution: * - ShardCommandExecution:
* Execution of a Task across a list of placements. * Execution of a Task across a list of placements.
* - TaskPlacementExecution: * - TaskPlacementExecution:

View File

@ -268,7 +268,7 @@ RemoteFileDestReceiverStartup(DestReceiver *dest, int operation,
/* /*
* PrepareIntermediateResultBroadcast gets a RemoteFileDestReceiver and does * PrepareIntermediateResultBroadcast gets a RemoteFileDestReceiver and does
* the necessary initilizations including initiating the remote connnections * the necessary initilizations including initiating the remote connections
* and creating the local file, which is necessary (it might be both). * and creating the local file, which is necessary (it might be both).
*/ */
static void static void

View File

@ -101,7 +101,7 @@ typedef struct DependencyDefinition
* address is used for dependencies that are artificially added during the * address is used for dependencies that are artificially added during the
* chasing. Since they are added by citus code we assume the dependency needs to * chasing. Since they are added by citus code we assume the dependency needs to
* be chased anyway, ofcourse it will only actually be chased if the object is a * be chased anyway, ofcourse it will only actually be chased if the object is a
* suppported object by citus * supported object by citus
*/ */
ObjectAddress address; ObjectAddress address;
} data; } data;

View File

@ -40,7 +40,7 @@ static void ExecuteCommandsInParallelAndStoreResults(StringInfo *nodeNameArray,
StringInfo *commandStringArray, StringInfo *commandStringArray,
bool *statusArray, bool *statusArray,
StringInfo *resultStringArray, StringInfo *resultStringArray,
int commmandCount); int commandCount);
static bool GetConnectionStatusAndResult(MultiConnection *connection, bool *resultStatus, static bool GetConnectionStatusAndResult(MultiConnection *connection, bool *resultStatus,
StringInfo queryResultString); StringInfo queryResultString);
static bool EvaluateQueryResult(MultiConnection *connection, PGresult *queryResult, static bool EvaluateQueryResult(MultiConnection *connection, PGresult *queryResult,
@ -51,7 +51,7 @@ static void ExecuteCommandsAndStoreResults(StringInfo *nodeNameArray,
StringInfo *commandStringArray, StringInfo *commandStringArray,
bool *statusArray, bool *statusArray,
StringInfo *resultStringArray, StringInfo *resultStringArray,
int commmandCount); int commandCount);
static bool ExecuteRemoteQueryOrCommand(char *nodeName, uint32 nodePort, static bool ExecuteRemoteQueryOrCommand(char *nodeName, uint32 nodePort,
char *queryString, StringInfo queryResult); char *queryString, StringInfo queryResult);
static Tuplestorestate * CreateTupleStore(TupleDesc tupleDescriptor, static Tuplestorestate * CreateTupleStore(TupleDesc tupleDescriptor,
@ -216,14 +216,14 @@ static void
ExecuteCommandsInParallelAndStoreResults(StringInfo *nodeNameArray, int *nodePortArray, ExecuteCommandsInParallelAndStoreResults(StringInfo *nodeNameArray, int *nodePortArray,
StringInfo *commandStringArray, StringInfo *commandStringArray,
bool *statusArray, StringInfo *resultStringArray, bool *statusArray, StringInfo *resultStringArray,
int commmandCount) int commandCount)
{ {
MultiConnection **connectionArray = MultiConnection **connectionArray =
palloc0(commmandCount * sizeof(MultiConnection *)); palloc0(commandCount * sizeof(MultiConnection *));
int finishedCount = 0; int finishedCount = 0;
/* start connections asynchronously */ /* start connections asynchronously */
for (int commandIndex = 0; commandIndex < commmandCount; commandIndex++) for (int commandIndex = 0; commandIndex < commandCount; commandIndex++)
{ {
char *nodeName = nodeNameArray[commandIndex]->data; char *nodeName = nodeNameArray[commandIndex]->data;
int nodePort = nodePortArray[commandIndex]; int nodePort = nodePortArray[commandIndex];
@ -233,7 +233,7 @@ ExecuteCommandsInParallelAndStoreResults(StringInfo *nodeNameArray, int *nodePor
} }
/* establish connections */ /* establish connections */
for (int commandIndex = 0; commandIndex < commmandCount; commandIndex++) for (int commandIndex = 0; commandIndex < commandCount; commandIndex++)
{ {
MultiConnection *connection = connectionArray[commandIndex]; MultiConnection *connection = connectionArray[commandIndex];
StringInfo queryResultString = resultStringArray[commandIndex]; StringInfo queryResultString = resultStringArray[commandIndex];
@ -257,7 +257,7 @@ ExecuteCommandsInParallelAndStoreResults(StringInfo *nodeNameArray, int *nodePor
} }
/* send queries at once */ /* send queries at once */
for (int commandIndex = 0; commandIndex < commmandCount; commandIndex++) for (int commandIndex = 0; commandIndex < commandCount; commandIndex++)
{ {
MultiConnection *connection = connectionArray[commandIndex]; MultiConnection *connection = connectionArray[commandIndex];
char *queryString = commandStringArray[commandIndex]->data; char *queryString = commandStringArray[commandIndex]->data;
@ -284,9 +284,9 @@ ExecuteCommandsInParallelAndStoreResults(StringInfo *nodeNameArray, int *nodePor
} }
/* check for query results */ /* check for query results */
while (finishedCount < commmandCount) while (finishedCount < commandCount)
{ {
for (int commandIndex = 0; commandIndex < commmandCount; commandIndex++) for (int commandIndex = 0; commandIndex < commandCount; commandIndex++)
{ {
MultiConnection *connection = connectionArray[commandIndex]; MultiConnection *connection = connectionArray[commandIndex];
StringInfo queryResultString = resultStringArray[commandIndex]; StringInfo queryResultString = resultStringArray[commandIndex];
@ -311,7 +311,7 @@ ExecuteCommandsInParallelAndStoreResults(StringInfo *nodeNameArray, int *nodePor
CHECK_FOR_INTERRUPTS(); CHECK_FOR_INTERRUPTS();
if (finishedCount < commmandCount) if (finishedCount < commandCount)
{ {
long sleepIntervalPerCycle = RemoteTaskCheckInterval * 1000L; long sleepIntervalPerCycle = RemoteTaskCheckInterval * 1000L;
pg_usleep(sleepIntervalPerCycle); pg_usleep(sleepIntervalPerCycle);
@ -468,9 +468,9 @@ StoreErrorMessage(MultiConnection *connection, StringInfo queryResultString)
static void static void
ExecuteCommandsAndStoreResults(StringInfo *nodeNameArray, int *nodePortArray, ExecuteCommandsAndStoreResults(StringInfo *nodeNameArray, int *nodePortArray,
StringInfo *commandStringArray, bool *statusArray, StringInfo *commandStringArray, bool *statusArray,
StringInfo *resultStringArray, int commmandCount) StringInfo *resultStringArray, int commandCount)
{ {
for (int commandIndex = 0; commandIndex < commmandCount; commandIndex++) for (int commandIndex = 0; commandIndex < commandCount; commandIndex++)
{ {
char *nodeName = nodeNameArray[commandIndex]->data; char *nodeName = nodeNameArray[commandIndex]->data;
int32 nodePort = nodePortArray[commandIndex]; int32 nodePort = nodePortArray[commandIndex];

View File

@ -331,7 +331,7 @@ CreateAllTargetListForRelation(Oid relationId, List *requiredAttributes)
/* /*
* For dropped columns, we generate a dummy null column because * For dropped columns, we generate a dummy null column because
* varattno in relation and subquery are different things, however if * varattno in relation and subquery are different things, however if
* we put the NULL columns to the subquery for the droppped columns, * we put the NULL columns to the subquery for the dropped columns,
* they will point to the same variable. * they will point to the same variable.
*/ */
TargetEntry *nullTargetEntry = CreateUnusedTargetEntry(resNo); TargetEntry *nullTargetEntry = CreateUnusedTargetEntry(resNo);

View File

@ -64,7 +64,7 @@
* - All multi-shard modifications (DDLs, COPY, UPDATE, DELETE, INSERT .. SELECT) * - All multi-shard modifications (DDLs, COPY, UPDATE, DELETE, INSERT .. SELECT)
* - All multi-shard queries with CTEs (modifying CTEs, read-only CTEs) * - All multi-shard queries with CTEs (modifying CTEs, read-only CTEs)
* - All recursively planned subqueries * - All recursively planned subqueries
* - All queries within transaction blocks (BEGIN; query; COMMMIT;) * - All queries within transaction blocks (BEGIN; query; COMMIT;)
* *
* In other words, the following types of queries won't be observed in these * In other words, the following types of queries won't be observed in these
* views: * views:

View File

@ -6,7 +6,7 @@
* are intended to track the relation accesses within a transaction. The * are intended to track the relation accesses within a transaction. The
* logic here is mostly useful when a reference table is referred by * logic here is mostly useful when a reference table is referred by
* a distributed table via a foreign key. Whenever such a pair of tables * a distributed table via a foreign key. Whenever such a pair of tables
* are acccesed inside a transaction, Citus should detect and act * are accessed inside a transaction, Citus should detect and act
* accordingly. * accordingly.
* *
* Copyright (c) Citus Data, Inc. * Copyright (c) Citus Data, Inc.
@ -798,7 +798,7 @@ CheckConflictingRelationAccesses(Oid relationId, ShardPlacementAccessType access
/* /*
* Switching to sequential mode is admittedly confusing and, could be useless * Switching to sequential mode is admittedly confusing and, could be useless
* and less performant in some cases. However, if we do not switch to * and less performant in some cases. However, if we do not switch to
* sequential mode at this point, we'd loose the opportunity to do so * sequential mode at this point, we'd lose the opportunity to do so
* later when a parallel query is executed on the hash distributed relations * later when a parallel query is executed on the hash distributed relations
* that are referencing this reference table. * that are referencing this reference table.
*/ */

View File

@ -3,7 +3,7 @@
* worker_merge_protocol.c * worker_merge_protocol.c
* *
* Routines for merging partitioned files into a single file or table. Merging * Routines for merging partitioned files into a single file or table. Merging
* files is one of the threee distributed execution primitives that we apply on * files is one of the three distributed execution primitives that we apply on
* worker nodes. * worker nodes.
* *
* Copyright (c) Citus Data, Inc. * Copyright (c) Citus Data, Inc.

View File

@ -17,10 +17,10 @@ BEGIN
END; END;
$$LANGUAGE plpgsql; $$LANGUAGE plpgsql;
-- Create a function to ignore worker plans in explain output -- Create a function to ignore worker plans in explain output
CREATE OR REPLACE FUNCTION coordinator_plan(explain_commmand text, out query_plan text) CREATE OR REPLACE FUNCTION coordinator_plan(explain_command text, out query_plan text)
RETURNS SETOF TEXT AS $$ RETURNS SETOF TEXT AS $$
BEGIN BEGIN
FOR query_plan IN execute explain_commmand LOOP FOR query_plan IN execute explain_command LOOP
RETURN next; RETURN next;
IF query_plan LIKE '%Task Count:%' IF query_plan LIKE '%Task Count:%'
THEN THEN
@ -30,12 +30,12 @@ BEGIN
RETURN; RETURN;
END; $$ language plpgsql; END; $$ language plpgsql;
-- helper function that returns true if output of given explain has "is not null" (case in-sensitive) -- helper function that returns true if output of given explain has "is not null" (case in-sensitive)
CREATE OR REPLACE FUNCTION explain_has_is_not_null(explain_commmand text) CREATE OR REPLACE FUNCTION explain_has_is_not_null(explain_command text)
RETURNS BOOLEAN AS $$ RETURNS BOOLEAN AS $$
DECLARE DECLARE
query_plan text; query_plan text;
BEGIN BEGIN
FOR query_plan IN EXECUTE explain_commmand LOOP FOR query_plan IN EXECUTE explain_command LOOP
IF query_plan ILIKE '%is not null%' IF query_plan ILIKE '%is not null%'
THEN THEN
RETURN true; RETURN true;
@ -44,12 +44,12 @@ BEGIN
RETURN false; RETURN false;
END; $$ language plpgsql; END; $$ language plpgsql;
-- helper function that returns true if output of given explain has "is not null" (case in-sensitive) -- helper function that returns true if output of given explain has "is not null" (case in-sensitive)
CREATE OR REPLACE FUNCTION explain_has_distributed_subplan(explain_commmand text) CREATE OR REPLACE FUNCTION explain_has_distributed_subplan(explain_command text)
RETURNS BOOLEAN AS $$ RETURNS BOOLEAN AS $$
DECLARE DECLARE
query_plan text; query_plan text;
BEGIN BEGIN
FOR query_plan IN EXECUTE explain_commmand LOOP FOR query_plan IN EXECUTE explain_command LOOP
IF query_plan ILIKE '%Distributed Subplan %_%' IF query_plan ILIKE '%Distributed Subplan %_%'
THEN THEN
RETURN true; RETURN true;

View File

@ -20,10 +20,10 @@ END;
$$LANGUAGE plpgsql; $$LANGUAGE plpgsql;
-- Create a function to ignore worker plans in explain output -- Create a function to ignore worker plans in explain output
CREATE OR REPLACE FUNCTION coordinator_plan(explain_commmand text, out query_plan text) CREATE OR REPLACE FUNCTION coordinator_plan(explain_command text, out query_plan text)
RETURNS SETOF TEXT AS $$ RETURNS SETOF TEXT AS $$
BEGIN BEGIN
FOR query_plan IN execute explain_commmand LOOP FOR query_plan IN execute explain_command LOOP
RETURN next; RETURN next;
IF query_plan LIKE '%Task Count:%' IF query_plan LIKE '%Task Count:%'
THEN THEN
@ -34,12 +34,12 @@ BEGIN
END; $$ language plpgsql; END; $$ language plpgsql;
-- helper function that returns true if output of given explain has "is not null" (case in-sensitive) -- helper function that returns true if output of given explain has "is not null" (case in-sensitive)
CREATE OR REPLACE FUNCTION explain_has_is_not_null(explain_commmand text) CREATE OR REPLACE FUNCTION explain_has_is_not_null(explain_command text)
RETURNS BOOLEAN AS $$ RETURNS BOOLEAN AS $$
DECLARE DECLARE
query_plan text; query_plan text;
BEGIN BEGIN
FOR query_plan IN EXECUTE explain_commmand LOOP FOR query_plan IN EXECUTE explain_command LOOP
IF query_plan ILIKE '%is not null%' IF query_plan ILIKE '%is not null%'
THEN THEN
RETURN true; RETURN true;
@ -49,12 +49,12 @@ BEGIN
END; $$ language plpgsql; END; $$ language plpgsql;
-- helper function that returns true if output of given explain has "is not null" (case in-sensitive) -- helper function that returns true if output of given explain has "is not null" (case in-sensitive)
CREATE OR REPLACE FUNCTION explain_has_distributed_subplan(explain_commmand text) CREATE OR REPLACE FUNCTION explain_has_distributed_subplan(explain_command text)
RETURNS BOOLEAN AS $$ RETURNS BOOLEAN AS $$
DECLARE DECLARE
query_plan text; query_plan text;
BEGIN BEGIN
FOR query_plan IN EXECUTE explain_commmand LOOP FOR query_plan IN EXECUTE explain_command LOOP
IF query_plan ILIKE '%Distributed Subplan %_%' IF query_plan ILIKE '%Distributed Subplan %_%'
THEN THEN
RETURN true; RETURN true;