diff --git a/src/backend/distributed/operations/repair_shards.c b/src/backend/distributed/operations/repair_shards.c index 815ecce91..bf856a997 100644 --- a/src/backend/distributed/operations/repair_shards.c +++ b/src/backend/distributed/operations/repair_shards.c @@ -1182,9 +1182,9 @@ CopyShardTablesViaLogicalReplication(List *shardIntervalList, char *sourceNodeNa int32 targetNodePort) { AcquireLogicalReplicationLock(); -elog(WARNING,"before drop leftovers"); +elog(LOG,"before drop leftovers"); DropAllLogicalReplicationLeftovers(SHARD_MOVE); -elog(WARNING,"after drop leftovers"); +elog(LOG,"after drop leftovers"); MemoryContext localContext = AllocSetContextCreate(CurrentMemoryContext, "CopyShardTablesViaLogicalReplication", diff --git a/src/backend/distributed/replication/multi_logical_replication.c b/src/backend/distributed/replication/multi_logical_replication.c index 26fe9af38..bfedb998b 100644 --- a/src/backend/distributed/replication/multi_logical_replication.c +++ b/src/backend/distributed/replication/multi_logical_replication.c @@ -320,10 +320,10 @@ LogicallyReplicateShards(List *shardList, char *sourceNodeName, int sourceNodePo * the constraints earlier. */ CreateForeignConstraintsToReferenceTable(logicalRepTargetList); - elog(WARNING,"before drop subs - try"); + elog(LOG,"before drop subs - try"); /* we're done, cleanup the publication and subscription */ DropSubscriptions(logicalRepTargetList); - elog(WARNING,"after drop subs - try"); + elog(LOG,"after drop subs - try"); DropReplicationSlots(sourceConnection, logicalRepTargetList); DropPublications(sourceConnection, publicationInfoHash); @@ -349,9 +349,9 @@ LogicallyReplicateShards(List *shardList, char *sourceNodeName, int sourceNodePo /* reconnect if the connection failed or is waiting for a command */ // RecreateGroupedLogicalRepTargetsConnections(groupedLogicalRepTargetsHash, // superUser, databaseName); - elog(WARNING,"before drop subs - catch"); + elog(LOG,"before drop subs - catch"); DropSubscriptions(logicalRepTargetList); - elog(WARNING,"after drop subs - catch"); + elog(LOG,"after drop subs - catch"); /* reconnect if the connection failed or is waiting for a command */ if (PQstatus(sourceConnection->pgConn) != CONNECTION_OK || PQisBusy(sourceConnection->pgConn)) diff --git a/src/test/regress/sql/failure_online_move_shard_placement.sql b/src/test/regress/sql/failure_online_move_shard_placement.sql index 01339c7c7..9388d4ebc 100644 --- a/src/test/regress/sql/failure_online_move_shard_placement.sql +++ b/src/test/regress/sql/failure_online_move_shard_placement.sql @@ -89,11 +89,14 @@ SELECT master_move_shard_placement(101, 'localhost', :worker_1_port, 'localhost' SELECT citus.mitmproxy('conn.onQuery(query="^ALTER SUBSCRIPTION .* DISABLE").cancel(' || :pid || ')'); SELECT master_move_shard_placement(101, 'localhost', :worker_1_port, 'localhost', :worker_2_proxy_port); +SET client_min_messages TO LOG; -- failure on dropping subscription SELECT citus.mitmproxy('conn.onQuery(query="^ALTER SUBSCRIPTION").after(2).kill()'); SELECT master_move_shard_placement(101, 'localhost', :worker_1_port, 'localhost', :worker_2_proxy_port); +-- try again SELECT citus.mitmproxy('conn.onQuery(query="nonexistingquery").kill()'); SELECT master_move_shard_placement(101, 'localhost', :worker_1_port, 'localhost', :worker_2_proxy_port); +RESET client_min_messages; -- cancellation on dropping subscription SELECT citus.mitmproxy('conn.onQuery(query="^DROP SUBSCRIPTION").cancel(' || :pid || ')');