/*------------------------------------------------------------------------- * * delete_protocol.c * * Routine for deleting shards in the distributed cluster. This function takes * in a delete command and deletes a shard if and only if all rows in the shard * satisfy the conditions in the delete command. * * Copyright (c) Citus Data, Inc. * * $Id$ * *------------------------------------------------------------------------- */ #include "postgres.h" #include "distributed/pg_version_constants.h" #include "c.h" #include "fmgr.h" #include "libpq-fe.h" #include "miscadmin.h" #include "port.h" #include #include "access/xact.h" #include "catalog/namespace.h" #include "commands/dbcommands.h" #include "distributed/commands/utility_hook.h" #include "distributed/connection_management.h" #include "distributed/deparse_shard_query.h" #include "distributed/listutils.h" #include "distributed/local_executor.h" #include "distributed/coordinator_protocol.h" #include "distributed/metadata_sync.h" #include "distributed/multi_client_executor.h" #include "distributed/multi_join_order.h" #include "distributed/multi_logical_planner.h" #include "distributed/multi_physical_planner.h" #include "distributed/multi_server_executor.h" #include "distributed/pg_dist_partition.h" #include "distributed/pg_dist_shard.h" #include "distributed/placement_connection.h" #include "distributed/relay_utility.h" #include "distributed/remote_commands.h" #include "distributed/worker_protocol.h" #include "distributed/worker_transaction.h" #include "lib/stringinfo.h" #include "nodes/nodeFuncs.h" #include "nodes/nodes.h" #include "nodes/parsenodes.h" #include "nodes/pg_list.h" #include "nodes/primnodes.h" #include "optimizer/clauses.h" #include "nodes/pathnodes.h" #include "optimizer/optimizer.h" #include "optimizer/restrictinfo.h" #include "storage/lock.h" #include "storage/lmgr.h" #include "tcop/tcopprot.h" #include "utils/array.h" #include "utils/builtins.h" #include "utils/elog.h" #include "utils/errcodes.h" #include "utils/lsyscache.h" #include "utils/varlena.h" /* Local functions forward declarations */ static void CheckTableCount(Query *deleteQuery); static void CheckDeleteCriteria(Node *deleteCriteria); static void CheckPartitionColumn(Oid relationId, Node *whereClause); static List * ShardsMatchingDeleteCriteria(Oid relationId, List *shardList, Node *deleteCriteria); static int DropShards(Oid relationId, char *schemaName, char *relationName, List *deletableShardIntervalList); static List * DropTaskList(Oid relationId, char *schemaName, char *relationName, List *deletableShardIntervalList); static void ExecuteDropShardPlacementCommandRemotely(ShardPlacement *shardPlacement, const char *shardRelationName, const char *dropShardPlacementCommand); static char * CreateDropShardPlacementCommand(const char *schemaName, const char *shardRelationName, char storageType); /* exports for SQL callable functions */ PG_FUNCTION_INFO_V1(master_apply_delete_command); PG_FUNCTION_INFO_V1(citus_drop_all_shards); PG_FUNCTION_INFO_V1(master_drop_all_shards); PG_FUNCTION_INFO_V1(master_drop_sequences); /* * master_apply_delete_command takes in a delete command, finds shards that * match the criteria defined in the delete command, drops the found shards from * the worker nodes, and updates the corresponding metadata on the master node. * This function drops a shard if and only if all rows in the shard satisfy * the conditions in the delete command. Note that this function only accepts * conditions on the partition key and if no condition is provided then all * shards are deleted. * * We mark shard placements that we couldn't drop as to be deleted later. If a * shard satisfies the given conditions, we delete it from shard metadata table * even though related shard placements are not deleted. */ Datum master_apply_delete_command(PG_FUNCTION_ARGS) { CheckCitusVersion(ERROR); EnsureCoordinator(); text *queryText = PG_GETARG_TEXT_P(0); char *queryString = text_to_cstring(queryText); List *deletableShardIntervalList = NIL; bool failOK = false; RawStmt *rawStmt = (RawStmt *) ParseTreeRawStmt(queryString); Node *queryTreeNode = rawStmt->stmt; if (!IsA(queryTreeNode, DeleteStmt)) { ereport(ERROR, (errmsg("query \"%s\" is not a delete statement", ApplyLogRedaction(queryString)))); } DeleteStmt *deleteStatement = (DeleteStmt *) queryTreeNode; char *schemaName = deleteStatement->relation->schemaname; char *relationName = deleteStatement->relation->relname; /* * We take an exclusive lock while dropping shards to prevent concurrent * writes. We don't want to block SELECTs, which means queries might fail * if they access a shard that has just been dropped. */ LOCKMODE lockMode = ExclusiveLock; Oid relationId = RangeVarGetRelid(deleteStatement->relation, lockMode, failOK); /* schema-prefix if it is not specified already */ if (schemaName == NULL) { Oid schemaId = get_rel_namespace(relationId); schemaName = get_namespace_name(schemaId); } CheckDistributedTable(relationId); EnsureTablePermissions(relationId, ACL_DELETE); List *queryTreeList = pg_analyze_and_rewrite(rawStmt, queryString, NULL, 0, NULL); Query *deleteQuery = (Query *) linitial(queryTreeList); CheckTableCount(deleteQuery); /* get where clause and flatten it */ Node *whereClause = (Node *) deleteQuery->jointree->quals; Node *deleteCriteria = eval_const_expressions(NULL, whereClause); if (IsCitusTableType(relationId, HASH_DISTRIBUTED)) { ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot delete from hash distributed table with this " "command"), errdetail("Delete statements on hash-partitioned tables " "are not supported with master_apply_delete_command."), errhint("Use the DELETE command instead."))); } else if (IsCitusTableType(relationId, CITUS_TABLE_WITH_NO_DIST_KEY)) { ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot delete from table"), errdetail("Delete statements on reference and " "local tables are not supported."))); } CheckDeleteCriteria(deleteCriteria); CheckPartitionColumn(relationId, deleteCriteria); List *shardIntervalList = LoadShardIntervalList(relationId); /* drop all shards if where clause is not present */ if (deleteCriteria == NULL) { deletableShardIntervalList = shardIntervalList; ereport(DEBUG2, (errmsg("dropping all shards for \"%s\"", relationName))); } else { deletableShardIntervalList = ShardsMatchingDeleteCriteria(relationId, shardIntervalList, deleteCriteria); } int droppedShardCount = DropShards(relationId, schemaName, relationName, deletableShardIntervalList); PG_RETURN_INT32(droppedShardCount); } /* * citus_drop_all_shards attempts to drop all shards for a given relation. * Unlike master_apply_delete_command, this function can be called even * if the table has already been dropped. */ Datum citus_drop_all_shards(PG_FUNCTION_ARGS) { CheckCitusVersion(ERROR); Oid relationId = PG_GETARG_OID(0); text *schemaNameText = PG_GETARG_TEXT_P(1); text *relationNameText = PG_GETARG_TEXT_P(2); char *schemaName = text_to_cstring(schemaNameText); char *relationName = text_to_cstring(relationNameText); /* * The SQL_DROP trigger calls this function even for tables that are * not distributed. In that case, silently ignore and return -1. */ if (!IsCitusTable(relationId) || !EnableDDLPropagation) { PG_RETURN_INT32(-1); } EnsureCoordinator(); CheckTableSchemaNameForDrop(relationId, &schemaName, &relationName); /* * citus_drop_all_shards is typically called from the DROP TABLE trigger, * but could be called by a user directly. Make sure we have an * AccessExclusiveLock to prevent any other commands from running on this table * concurrently. */ LockRelationOid(relationId, AccessExclusiveLock); List *shardIntervalList = LoadShardIntervalList(relationId); int droppedShardCount = DropShards(relationId, schemaName, relationName, shardIntervalList); PG_RETURN_INT32(droppedShardCount); } /* * master_drop_all_shards is a wrapper function for old UDF name. */ Datum master_drop_all_shards(PG_FUNCTION_ARGS) { return citus_drop_all_shards(fcinfo); } /* * master_drop_sequences was previously used to drop sequences on workers * when using metadata syncing. * * It may still be called when dropping objects during CREATE EXTENSION, * hence the function remains in place. */ Datum master_drop_sequences(PG_FUNCTION_ARGS) { PG_RETURN_VOID(); } /* * CheckTableSchemaNameForDrop errors out if the current user does not * have permission to un-distribute the given relation, taking into * account that it may be called from the drop trigger. If the table exists, * the function rewrites the given table and schema name. */ void CheckTableSchemaNameForDrop(Oid relationId, char **schemaName, char **tableName) { char *tempTableName = get_rel_name(relationId); if (tempTableName != NULL) { /* ensure proper values are used if the table exists */ Oid schemaId = get_rel_namespace(relationId); (*schemaName) = get_namespace_name(schemaId); (*tableName) = tempTableName; EnsureTableOwner(relationId); } } /* * DropShards drops all given shards in a relation. The id, name and schema * for the relation are explicitly provided, since this function may be * called when the table is already dropped. * * We mark shard placements that we couldn't drop as to be deleted later, but * we do delete the shard metadadata. */ static int DropShards(Oid relationId, char *schemaName, char *relationName, List *deletableShardIntervalList) { Assert(OidIsValid(relationId)); Assert(schemaName != NULL); Assert(relationName != NULL); UseCoordinatedTransaction(); /* * We will use below variable accross this function to decide if we can * use local execution */ int32 localGroupId = GetLocalGroupId(); /* DROP table commands are currently only supported from the coordinator */ Assert(localGroupId == COORDINATOR_GROUP_ID); /* * At this point we intentionally decided to not use 2PC for reference * tables */ if (MultiShardCommitProtocol == COMMIT_PROTOCOL_2PC) { Use2PCForCoordinatedTransaction(); } List *dropTaskList = DropTaskList(relationId, schemaName, relationName, deletableShardIntervalList); bool shouldExecuteTasksLocally = ShouldExecuteTasksLocally(dropTaskList); Task *task = NULL; foreach_ptr(task, dropTaskList) { uint64 shardId = task->anchorShardId; ShardPlacement *shardPlacement = NULL; foreach_ptr(shardPlacement, task->taskPlacementList) { uint64 shardPlacementId = shardPlacement->placementId; int32 shardPlacementGroupId = shardPlacement->groupId; bool isLocalShardPlacement = (shardPlacementGroupId == localGroupId); if (isLocalShardPlacement && DropSchemaOrDBInProgress() && localGroupId == COORDINATOR_GROUP_ID) { /* * The active DROP SCHEMA/DATABASE ... CASCADE will drop the * shard, if we try to drop it over another connection, we will * get into a distributed deadlock. Hence, just delete the shard * placement metadata and skip it for now. */ DeleteShardPlacementRow(shardPlacementId); continue; } /* * If it is a local placement of a distributed table or a reference table, * then execute the DROP command locally. */ if (isLocalShardPlacement && shouldExecuteTasksLocally) { List *singleTaskList = list_make1(task); ExecuteLocalUtilityTaskList(singleTaskList); } else { /* * Either it was not a local placement or we could not use * local execution even if it was a local placement. * If it is the second case, then it is possibly because in * current transaction, some commands or queries connected * to local group as well. * * Regardless of the node is a remote node or the current node, * try to open a new connection (or use an existing one) to * connect to that node to drop the shard placement over that * remote connection. */ const char *dropShardPlacementCommand = TaskQueryString(task); ExecuteDropShardPlacementCommandRemotely(shardPlacement, relationName, dropShardPlacementCommand); if (isLocalShardPlacement) { SetLocalExecutionStatus(LOCAL_EXECUTION_DISABLED); } } DeleteShardPlacementRow(shardPlacementId); } /* * Now that we deleted all placements of the shard (or their metadata), * delete the shard metadata as well. */ DeleteShardRow(shardId); } int droppedShardCount = list_length(deletableShardIntervalList); return droppedShardCount; } /* * DropTaskList returns a list of tasks to execute a DROP command on shard * placements of distributed table. This is handled separately from other * DDL commands because we handle it via the DROP trigger, which is called * whenever a drop cascades. */ static List * DropTaskList(Oid relationId, char *schemaName, char *relationName, List *deletableShardIntervalList) { /* resulting task list */ List *taskList = NIL; /* enumerate the tasks when putting them to the taskList */ int taskId = 1; ShardInterval *shardInterval = NULL; foreach_ptr(shardInterval, deletableShardIntervalList) { Assert(shardInterval->relationId == relationId); uint64 shardId = shardInterval->shardId; char storageType = shardInterval->storageType; char *shardRelationName = pstrdup(relationName); /* build shard relation name */ AppendShardIdToName(&shardRelationName, shardId); char *dropShardPlacementCommand = CreateDropShardPlacementCommand(schemaName, shardRelationName, storageType); Task *task = CitusMakeNode(Task); task->jobId = INVALID_JOB_ID; task->taskId = taskId++; task->taskType = DDL_TASK; SetTaskQueryString(task, dropShardPlacementCommand); task->dependentTaskList = NULL; task->replicationModel = REPLICATION_MODEL_INVALID; task->anchorShardId = shardId; task->taskPlacementList = ShardPlacementList(shardId); taskList = lappend(taskList, task); } return taskList; } /* * ExecuteDropShardPlacementCommandRemotely executes the given DROP shard command * via remote critical connection. */ static void ExecuteDropShardPlacementCommandRemotely(ShardPlacement *shardPlacement, const char *relationName, const char *dropShardPlacementCommand) { Assert(shardPlacement != NULL); Assert(relationName != NULL); Assert(dropShardPlacementCommand != NULL); uint32 connectionFlags = FOR_DDL; MultiConnection *connection = GetPlacementConnection(connectionFlags, shardPlacement, NULL); /* * This code-path doesn't support optional connections, so we don't expect * NULL connections. */ Assert(connection != NULL); RemoteTransactionBeginIfNecessary(connection); if (PQstatus(connection->pgConn) != CONNECTION_OK) { uint64 placementId = shardPlacement->placementId; char *workerName = shardPlacement->nodeName; uint32 workerPort = shardPlacement->nodePort; /* build shard relation name */ uint64 shardId = shardPlacement->shardId; char *shardRelationName = pstrdup(relationName); AppendShardIdToName(&shardRelationName, shardId); ereport(WARNING, (errmsg("could not connect to shard \"%s\" on node " "\"%s:%u\"", shardRelationName, workerName, workerPort), errdetail("Marking this shard placement for " "deletion"))); UpdateShardPlacementState(placementId, SHARD_STATE_TO_DELETE); return; } MarkRemoteTransactionCritical(connection); ExecuteCriticalRemoteCommand(connection, dropShardPlacementCommand); } /* * CreateDropShardPlacementCommand function builds the DROP command to drop * the given shard relation by qualifying it with schema name according to * shard relation's storage type. */ static char * CreateDropShardPlacementCommand(const char *schemaName, const char *shardRelationName, char storageType) { Assert(schemaName != NULL); Assert(shardRelationName != NULL); StringInfo workerDropQuery = makeStringInfo(); const char *quotedShardName = quote_qualified_identifier(schemaName, shardRelationName); /* build workerDropQuery according to shard storage type */ if (storageType == SHARD_STORAGE_TABLE) { appendStringInfo(workerDropQuery, DROP_REGULAR_TABLE_COMMAND, quotedShardName); } else if (storageType == SHARD_STORAGE_COLUMNAR || storageType == SHARD_STORAGE_FOREIGN) { appendStringInfo(workerDropQuery, DROP_FOREIGN_TABLE_COMMAND, quotedShardName); } else { /* no other storage type is expected here */ Assert(false); } return workerDropQuery->data; } /* Checks that delete is only on one table. */ static void CheckTableCount(Query *deleteQuery) { int rangeTableCount = list_length(deleteQuery->rtable); if (rangeTableCount > 1) { ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot delete from distributed table"), errdetail("Delete on multiple tables is not supported"))); } } /* Checks that delete criteria only consists of simple operator expressions. */ static void CheckDeleteCriteria(Node *deleteCriteria) { bool simpleOpExpression = true; if (deleteCriteria == NULL) { return; } if (is_opclause(deleteCriteria)) { simpleOpExpression = SimpleOpExpression((Expr *) deleteCriteria); } else if (IsA(deleteCriteria, BoolExpr)) { BoolExpr *deleteCriteriaExpression = (BoolExpr *) deleteCriteria; List *opExpressionList = deleteCriteriaExpression->args; Expr *opExpression = NULL; foreach_ptr(opExpression, opExpressionList) { if (!SimpleOpExpression(opExpression)) { simpleOpExpression = false; break; } } } else { simpleOpExpression = false; } if (!simpleOpExpression) { ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot delete from distributed table"), errdetail("Delete query has a complex operator expression"))); } } /* * CheckPartitionColumn checks that the given where clause is based only on the * partition key of the given relation id. */ static void CheckPartitionColumn(Oid relationId, Node *whereClause) { Var *partitionColumn = DistPartitionKeyOrError(relationId); List *columnList = pull_var_clause_default(whereClause); Var *var = NULL; foreach_ptr(var, columnList) { if (var->varattno != partitionColumn->varattno) { ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot delete from distributed table"), errdetail("Where clause includes a column other than " "partition column"))); } } } /* * ShardsMatchingDeleteCriteria selects shards to be deleted from the shard * interval list based on the delete criteria, and returns selected shards in * another list. We add a shard to the list if and only if all rows in the shard * satisfy the delete criteria. Note that this function does not expect * deleteCriteria to be NULL. */ static List * ShardsMatchingDeleteCriteria(Oid relationId, List *shardIntervalList, Node *deleteCriteria) { List *dropShardIntervalList = NIL; /* build the base expression for constraint */ Index rangeTableIndex = 1; Var *partitionColumn = PartitionColumn(relationId, rangeTableIndex); Node *baseConstraint = BuildBaseConstraint(partitionColumn); Assert(deleteCriteria != NULL); List *deleteCriteriaList = list_make1(deleteCriteria); /* walk over shard list and check if shards can be dropped */ ShardInterval *shardInterval = NULL; foreach_ptr(shardInterval, shardIntervalList) { if (shardInterval->minValueExists && shardInterval->maxValueExists) { List *restrictInfoList = NIL; /* set the min/max values in the base constraint */ UpdateConstraint(baseConstraint, shardInterval); BoolExpr *andExpr = (BoolExpr *) baseConstraint; Expr *lessThanExpr = (Expr *) linitial(andExpr->args); Expr *greaterThanExpr = (Expr *) lsecond(andExpr->args); RestrictInfo *lessThanRestrictInfo = make_simple_restrictinfo(lessThanExpr); RestrictInfo *greaterThanRestrictInfo = make_simple_restrictinfo( greaterThanExpr); restrictInfoList = lappend(restrictInfoList, lessThanRestrictInfo); restrictInfoList = lappend(restrictInfoList, greaterThanRestrictInfo); bool dropShard = predicate_implied_by(deleteCriteriaList, restrictInfoList, false); if (dropShard) { dropShardIntervalList = lappend(dropShardIntervalList, shardInterval); ereport(DEBUG2, (errmsg("delete criteria includes shardId " UINT64_FORMAT, shardInterval->shardId))); } } } return dropShardIntervalList; }