fix style

pull/3850/head
Sait Talha Nisanci 2020-05-19 16:04:43 +03:00
parent 3cb8d467b7
commit b0b3cb9e5b
5 changed files with 9 additions and 7 deletions

View File

@ -556,6 +556,7 @@ AdaptiveExecutorCreateScan(CustomScan *scan)
return (Node *) scanState; return (Node *) scanState;
} }
/* /*
* CoordinatorInsertSelectCrateScan creates the scan state for executing * CoordinatorInsertSelectCrateScan creates the scan state for executing
* INSERT..SELECT into a distributed table via the coordinator. * INSERT..SELECT into a distributed table via the coordinator.

View File

@ -114,7 +114,7 @@ JobExecutorType(DistributedPlan *distributedPlan)
if (HasReplicatedDistributedTable(distributedPlan->relationIdList)) if (HasReplicatedDistributedTable(distributedPlan->relationIdList))
{ {
ereport(ERROR, (errmsg( ereport(ERROR, (errmsg(
"repartitioning with shard_replication_factor > 1 is not supported"))); "repartitioning with shard_replication_factor > 1 is not supported")));
} }
return MULTI_EXECUTOR_ADAPTIVE; return MULTI_EXECUTOR_ADAPTIVE;
} }
@ -164,6 +164,7 @@ HasReplicatedDistributedTable(List *relationOids)
return false; return false;
} }
/* /*
* RemoveJobDirectory gets automatically called at portal drop (end of query) or * RemoveJobDirectory gets automatically called at portal drop (end of query) or
* at transaction abort. The function removes the job directory and releases the * at transaction abort. The function removes the job directory and releases the
@ -178,6 +179,7 @@ RemoveJobDirectory(uint64 jobId)
ResourceOwnerForgetJobDirectory(CurrentResourceOwner, jobId); ResourceOwnerForgetJobDirectory(CurrentResourceOwner, jobId);
} }
/* /*
* CheckIfSizeLimitIsExceeded checks if the limit is exceeded by intermediate * CheckIfSizeLimitIsExceeded checks if the limit is exceeded by intermediate
* results, if there is any. * results, if there is any.

View File

@ -129,6 +129,7 @@ TaskAndExecutionList(List *jobTaskList)
return taskAndExecutionList; return taskAndExecutionList;
} }
/* /*
* TaskHashCreate allocates memory for a task hash, initializes an * TaskHashCreate allocates memory for a task hash, initializes an
* empty hash, and returns this hash. * empty hash, and returns this hash.

View File

@ -54,8 +54,7 @@ static void CreateTaskTable(StringInfo schemaName, StringInfo relationName,
List *columnNameList, List *columnTypeList); List *columnNameList, List *columnTypeList);
static void CopyTaskFilesFromDirectory(StringInfo schemaName, StringInfo relationName, static void CopyTaskFilesFromDirectory(StringInfo schemaName, StringInfo relationName,
StringInfo sourceDirectoryName, Oid userId); StringInfo sourceDirectoryName, Oid userId);
static void static void CreateJobSchema(StringInfo schemaName, char *schemaOwner);
CreateJobSchema(StringInfo schemaName, char *schemaOwner);
/* exports for SQL callable functions */ /* exports for SQL callable functions */
@ -89,6 +88,7 @@ worker_create_schema(PG_FUNCTION_ARGS)
PG_RETURN_VOID(); PG_RETURN_VOID();
} }
/* /*
* CreateJobSchema creates a job schema with the given schema name. Note that * CreateJobSchema creates a job schema with the given schema name. Note that
* this function ensures that our pg_ prefixed schema names can be created. * this function ensures that our pg_ prefixed schema names can be created.
@ -141,7 +141,6 @@ CreateJobSchema(StringInfo schemaName, char *schemaOwner)
} }
/* /*
* worker_repartition_cleanup removes the job directory and schema with the given job id . * worker_repartition_cleanup removes the job directory and schema with the given job id .
*/ */

View File

@ -31,7 +31,6 @@
");" ");"
/* Enumeration that represents distributed executor types */ /* Enumeration that represents distributed executor types */
typedef enum typedef enum
{ {