Merge pull request #1145 from citusdata/tenant_isolation_community

Tenant isolation
pull/1152/head
Metin Döşlü 2017-01-20 13:49:04 +02:00 committed by GitHub
commit ac7235bfaa
25 changed files with 279 additions and 101 deletions

View File

@ -9,7 +9,7 @@ EXTVERSIONS = 5.0 5.0-1 5.0-2 \
5.1-1 5.1-2 5.1-3 5.1-4 5.1-5 5.1-6 5.1-7 5.1-8 \
5.2-1 5.2-2 5.2-3 5.2-4 \
6.0-1 6.0-2 6.0-3 6.0-4 6.0-5 6.0-6 6.0-7 6.0-8 6.0-9 6.0-10 6.0-11 6.0-12 6.0-13 6.0-14 6.0-15 6.0-16 6.0-17 6.0-18 \
6.1-1 6.1-2 6.1-3 6.1-4 6.1-5 6.1-6 6.1-7 6.1-8 6.1-9 6.1-10 6.1-11 6.1-12 6.1-13 6.1-14 6.1-15 6.1-16
6.1-1 6.1-2 6.1-3 6.1-4 6.1-5 6.1-6 6.1-7 6.1-8 6.1-9 6.1-10 6.1-11 6.1-12 6.1-13 6.1-14 6.1-15 6.1-16 6.1-17
# All citus--*.sql files in the source directory
DATA = $(patsubst $(citus_abs_srcdir)/%.sql,%.sql,$(wildcard $(citus_abs_srcdir)/$(EXTENSION)--*--*.sql))
@ -127,6 +127,8 @@ $(EXTENSION)--6.1-15.sql: $(EXTENSION)--6.1-14.sql $(EXTENSION)--6.1-14--6.1-15.
cat $^ > $@
$(EXTENSION)--6.1-16.sql: $(EXTENSION)--6.1-15.sql $(EXTENSION)--6.1-15--6.1-16.sql
cat $^ > $@
$(EXTENSION)--6.1-17.sql: $(EXTENSION)--6.1-16.sql $(EXTENSION)--6.1-16--6.1-17.sql
cat $^ > $@
NO_PGXS = 1

View File

@ -0,0 +1,19 @@
/* citus--6.1-16--6.1-17.sql */
SET search_path = 'pg_catalog';
CREATE FUNCTION isolate_tenant_to_new_shard(table_name regclass, tenant_id "any", cascade_option text DEFAULT '')
RETURNS bigint
LANGUAGE C STRICT
AS 'MODULE_PATHNAME', $$isolate_tenant_to_new_shard$$;
COMMENT ON FUNCTION isolate_tenant_to_new_shard(table_name regclass, tenant_id "any", cascade_option text)
IS 'isolate a tenant to its own shard and return the new shard id';
CREATE FUNCTION worker_hash(value "any")
RETURNS integer
LANGUAGE C STRICT
AS 'MODULE_PATHNAME', $$worker_hash$$;
COMMENT ON FUNCTION worker_hash(value "any")
IS 'calculate hashed value and return it';
RESET search_path;

View File

@ -1,6 +1,6 @@
# Citus extension
comment = 'Citus distributed database'
default_version = '6.1-16'
default_version = '6.1-17'
module_pathname = '$libdir/citus'
relocatable = false
schema = pg_catalog

View File

@ -50,10 +50,6 @@
#include "utils/palloc.h"
/* local function forward declarations */
static text * IntegerToText(int32 value);
/* declarations for dynamic loading */
PG_FUNCTION_INFO_V1(master_create_worker_shards);
@ -418,7 +414,7 @@ CheckHashPartitionedTable(Oid distributedTableId)
/* Helper function to convert an integer value to a text type */
static text *
text *
IntegerToText(int32 value)
{
text *valueText = NULL;

View File

@ -45,7 +45,6 @@ static void RepairShardPlacement(int64 shardId, char *sourceNodeName,
static void EnsureShardCanBeRepaired(int64 shardId, char *sourceNodeName,
int32 sourceNodePort, char *targetNodeName,
int32 targetNodePort);
static char * ConstructQualifiedShardName(ShardInterval *shardInterval);
static List * RecreateTableDDLCommandList(Oid relationId);
/* declarations for dynamic loading */
@ -350,10 +349,8 @@ CopyShardForeignConstraintCommandList(ShardInterval *shardInterval)
/*
* ConstuctQualifiedShardName creates the fully qualified name string of the
* given shard in <schema>.<table_name>_<shard_id> format.
*
* FIXME: Copied from Citus-MX, should be removed once those changes checked-in to Citus.
*/
static char *
char *
ConstructQualifiedShardName(ShardInterval *shardInterval)
{
Oid schemaId = get_rel_namespace(shardInterval->relationId);

View File

@ -0,0 +1,89 @@
/*-------------------------------------------------------------------------
*
* master_split_shards.c
*
* This file contains functions to split a shard according to a given
* distribution column value.
*
* Copyright (c) 2014-2017, Citus Data, Inc.
*
*-------------------------------------------------------------------------
*/
#include "postgres.h"
#include "c.h"
#include "fmgr.h"
#include "catalog/pg_class.h"
#include "distributed/colocation_utils.h"
#include "distributed/connection_cache.h"
#include "distributed/master_protocol.h"
#include "distributed/metadata_cache.h"
#include "distributed/metadata_sync.h"
#include "distributed/multi_join_order.h"
#include "distributed/multi_router_planner.h"
#include "distributed/pg_dist_partition.h"
#include "distributed/pg_dist_shard.h"
#include "distributed/remote_commands.h"
#include "distributed/resource_lock.h"
#include "distributed/worker_manager.h"
#include "distributed/worker_protocol.h"
#include "distributed/worker_transaction.h"
#include "nodes/pg_list.h"
#include "storage/lock.h"
#include "utils/builtins.h"
#include "utils/elog.h"
#include "utils/errcodes.h"
#include "utils/lsyscache.h"
#include "utils/typcache.h"
/* declarations for dynamic loading */
PG_FUNCTION_INFO_V1(isolate_tenant_to_new_shard);
PG_FUNCTION_INFO_V1(worker_hash);
/*
* isolate_tenant_to_new_shard isolates a tenant to its own shard by spliting
* the current matching shard.
*/
Datum
isolate_tenant_to_new_shard(PG_FUNCTION_ARGS)
{
ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
errmsg("isolate_tenant_to_new_shard() is only supported on "
"Citus Enterprise")));
}
/*
* worker_hash returns the hashed value of the given value.
*/
Datum
worker_hash(PG_FUNCTION_ARGS)
{
Datum valueDatum = PG_GETARG_DATUM(0);
Datum hashedValueDatum = 0;
TypeCacheEntry *typeEntry = NULL;
FmgrInfo *hashFunction = NULL;
Oid valueDataType = InvalidOid;
/* figure out hash function from the data type */
valueDataType = get_fn_expr_argtype(fcinfo->flinfo, 0);
typeEntry = lookup_type_cache(valueDataType, TYPECACHE_HASH_PROC_FINFO);
if (typeEntry->hash_proc_finfo.fn_oid == InvalidOid)
{
ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
errmsg("cannot find a hash function for the input type"),
errhint("Cast input to a data type with a hash function.")));
}
hashFunction = palloc0(sizeof(FmgrInfo));
fmgr_info_copy(hashFunction, &(typeEntry->hash_proc_finfo), CurrentMemoryContext);
/* calculate hash value */
hashedValueDatum = FunctionCall1(hashFunction, valueDatum);
PG_RETURN_INT32(hashedValueDatum);
}

View File

@ -513,7 +513,7 @@ TableOwnerResetCommand(Oid relationId)
/*
* ShardListInsertCommand generates a singe command that can be
* ShardListInsertCommand generates a single command that can be
* executed to replicate shard and shard placement metadata for the
* given shard intervals. The function assumes that each shard has a
* single placement, and asserts this information.
@ -634,6 +634,37 @@ ShardListInsertCommand(List *shardIntervalList)
}
/*
* ShardListDeleteCommand generates a command list that can be executed to delete
* shard and shard placement metadata for the given shard.
*/
List *
ShardDeleteCommandList(ShardInterval *shardInterval)
{
uint64 shardId = shardInterval->shardId;
List *commandList = NIL;
StringInfo deletePlacementCommand = NULL;
StringInfo deleteShardCommand = NULL;
/* create command to delete shard placements */
deletePlacementCommand = makeStringInfo();
appendStringInfo(deletePlacementCommand,
"DELETE FROM pg_dist_shard_placement WHERE shardid = %lu",
shardId);
commandList = lappend(commandList, deletePlacementCommand->data);
/* create command to delete shard */
deleteShardCommand = makeStringInfo();
appendStringInfo(deleteShardCommand,
"DELETE FROM pg_dist_shard WHERE shardid = %lu", shardId);
commandList = lappend(commandList, deleteShardCommand->data);
return commandList;
}
/*
* NodeDeleteCommand generate a command that can be
* executed to delete the metadata for a worker node.

View File

@ -95,8 +95,6 @@ static Task * RouterModifyTask(Query *originalQuery, Query *query);
static ShardInterval * TargetShardIntervalForModify(Query *query);
static List * QueryRestrictList(Query *query);
static bool FastShardPruningPossible(CmdType commandType, char partitionMethod);
static ShardInterval * FastShardPruning(Oid distributedTableId,
Const *partionColumnValue);
static Const * ExtractInsertPartitionValue(Query *query, Var *partitionColumn);
static Task * RouterSelectTask(Query *originalQuery,
RelationRestrictionContext *restrictionContext,
@ -1890,7 +1888,8 @@ TargetShardIntervalForModify(Query *query)
{
uint32 rangeTableId = 1;
Var *partitionColumn = PartitionColumn(distributedTableId, rangeTableId);
Const *partitionValue = ExtractInsertPartitionValue(query, partitionColumn);
Const *partitionValueConst = ExtractInsertPartitionValue(query, partitionColumn);
Datum partitionValue = partitionValueConst->constvalue;
ShardInterval *shardInterval = FastShardPruning(distributedTableId,
partitionValue);
@ -2001,8 +2000,8 @@ FastShardPruningPossible(CmdType commandType, char partitionMethod)
* the corresponding shard interval that the partitionValue should be in. FastShardPruning
* returns NULL if no ShardIntervals exist for the given partitionValue.
*/
static ShardInterval *
FastShardPruning(Oid distributedTableId, Const *partitionValue)
ShardInterval *
FastShardPruning(Oid distributedTableId, Datum partitionValue)
{
DistTableCacheEntry *cacheEntry = DistributedTableCacheEntry(distributedTableId);
int shardCount = cacheEntry->shardIntervalArrayLength;
@ -2030,9 +2029,8 @@ FastShardPruning(Oid distributedTableId, Const *partitionValue)
* Call FindShardInterval to find the corresponding shard interval for the
* given partition value.
*/
shardInterval = FindShardInterval(partitionValue->constvalue,
sortedShardIntervalArray, shardCount,
partitionMethod,
shardInterval = FindShardInterval(partitionValue, sortedShardIntervalArray,
shardCount, partitionMethod,
shardIntervalCompareFunction, hashFunction,
useBinarySearch);

View File

@ -28,10 +28,6 @@
#include "utils/lsyscache.h"
/* local function forward declarations */
static Datum ExtractIntegerDatum(char *input);
/* declarations for dynamic loading */
PG_FUNCTION_INFO_V1(initialize_remote_temp_table);
PG_FUNCTION_INFO_V1(count_remote_temp_table_rows);
@ -98,7 +94,7 @@ count_remote_temp_table_rows(PG_FUNCTION_ARGS)
else
{
char *countText = PQgetvalue(result, 0, 0);
count = ExtractIntegerDatum(countText);
count = StringToDatum(countText, INT4OID);
}
PQclear(result);
@ -189,24 +185,3 @@ set_connection_status_bad(PG_FUNCTION_ARGS)
PG_RETURN_BOOL(true);
}
/*
* ExtractIntegerDatum transforms an integer in textual form into a Datum.
*/
static Datum
ExtractIntegerDatum(char *input)
{
Oid typIoFunc = InvalidOid;
Oid typIoParam = InvalidOid;
Datum intDatum = 0;
FmgrInfo fmgrInfo;
memset(&fmgrInfo, 0, sizeof(fmgrInfo));
getTypeInputInfo(INT4OID, &typIoFunc, &typIoParam);
fmgr_info(typIoFunc, &fmgrInfo);
intDatum = InputFunctionCall(&fmgrInfo, input, typIoFunc, -1);
return intDatum;
}

View File

@ -841,7 +841,11 @@ ColocatedShardIntervalList(ShardInterval *shardInterval)
if ((partitionMethod == DISTRIBUTE_BY_APPEND) ||
(partitionMethod == DISTRIBUTE_BY_RANGE))
{
colocatedShardList = lappend(colocatedShardList, shardInterval);
ShardInterval *copyShardInterval = CitusMakeNode(ShardInterval);
CopyShardInterval(shardInterval, copyShardInterval);
colocatedShardList = lappend(colocatedShardList, copyShardInterval);
return colocatedShardList;
}
@ -857,6 +861,7 @@ ColocatedShardIntervalList(ShardInterval *shardInterval)
DistTableCacheEntry *colocatedTableCacheEntry =
DistributedTableCacheEntry(colocatedTableId);
ShardInterval *colocatedShardInterval = NULL;
ShardInterval *copyShardInterval = NULL;
/*
* Since we iterate over co-located tables, shard count of each table should be
@ -868,7 +873,10 @@ ColocatedShardIntervalList(ShardInterval *shardInterval)
colocatedShardInterval =
colocatedTableCacheEntry->sortedShardIntervalArray[shardIntervalIndex];
colocatedShardList = lappend(colocatedShardList, colocatedShardInterval);
copyShardInterval = CitusMakeNode(ShardInterval);
CopyShardInterval(colocatedShardInterval, copyShardInterval);
colocatedShardList = lappend(colocatedShardList, copyShardInterval);
}
Assert(list_length(colocatedTableList) == list_length(colocatedShardList));

View File

@ -30,6 +30,7 @@
#include "distributed/metadata_cache.h"
#include "distributed/metadata_sync.h"
#include "distributed/multi_join_order.h"
#include "distributed/multi_router_planner.h"
#include "distributed/pg_dist_node.h"
#include "distributed/reference_table_utils.h"
#include "distributed/shardinterval_utils.h"
@ -40,6 +41,7 @@
#include "storage/fd.h"
#include "utils/builtins.h"
#include "utils/fmgroids.h"
#include "utils/lsyscache.h"
#include "utils/rel.h"
#include "utils/relcache.h"
@ -179,20 +181,9 @@ master_initialize_node_metadata(PG_FUNCTION_ARGS)
Datum
get_shard_id_for_distribution_column(PG_FUNCTION_ARGS)
{
Oid relationId = InvalidOid;
Datum distributionValue = 0;
Var *distributionColumn = NULL;
char distributionMethod = 0;
Oid expectedElementType = InvalidOid;
Oid inputElementType = InvalidOid;
DistTableCacheEntry *cacheEntry = NULL;
int shardCount = 0;
ShardInterval **shardIntervalArray = NULL;
FmgrInfo *hashFunction = NULL;
FmgrInfo *compareFunction = NULL;
bool useBinarySearch = true;
ShardInterval *shardInterval = NULL;
char distributionMethod = 0;
Oid relationId = InvalidOid;
/*
* To have optional parameter as NULL, we defined this UDF as not strict, therefore
@ -227,6 +218,13 @@ get_shard_id_for_distribution_column(PG_FUNCTION_ARGS)
else if (distributionMethod == DISTRIBUTE_BY_HASH ||
distributionMethod == DISTRIBUTE_BY_RANGE)
{
Var *distributionColumn = NULL;
Oid distributionDataType = InvalidOid;
Oid inputDataType = InvalidOid;
char *distributionValueString = NULL;
Datum inputDatum = 0;
Datum distributionValueDatum = 0;
/* if given table is not reference table, distributionValue cannot be NULL */
if (PG_ARGISNULL(1))
{
@ -235,36 +233,17 @@ get_shard_id_for_distribution_column(PG_FUNCTION_ARGS)
"than reference tables.")));
}
distributionValue = PG_GETARG_DATUM(1);
inputDatum = PG_GETARG_DATUM(1);
inputDataType = get_fn_expr_argtype(fcinfo->flinfo, 1);
distributionValueString = DatumToString(inputDatum, inputDataType);
distributionColumn = PartitionKey(relationId);
expectedElementType = distributionColumn->vartype;
inputElementType = get_fn_expr_argtype(fcinfo->flinfo, 1);
if (expectedElementType != inputElementType)
{
ereport(ERROR, (errcode(ERRCODE_WRONG_OBJECT_TYPE),
errmsg("invalid distribution value type"),
errdetail("Type of the value does not match the type of the "
"distribution column. Expected type id: %d, given "
"type id: %d", expectedElementType,
inputElementType)));
}
distributionDataType = distributionColumn->vartype;
cacheEntry = DistributedTableCacheEntry(relationId);
distributionValueDatum = StringToDatum(distributionValueString,
distributionDataType);
if (distributionMethod == DISTRIBUTE_BY_HASH &&
cacheEntry->hasUniformHashDistribution)
{
useBinarySearch = false;
}
shardCount = cacheEntry->shardIntervalArrayLength;
shardIntervalArray = cacheEntry->sortedShardIntervalArray;
hashFunction = cacheEntry->hashFunction;
compareFunction = cacheEntry->shardIntervalCompareFunction;
shardInterval = FindShardInterval(distributionValue, shardIntervalArray,
shardCount, distributionMethod, compareFunction,
hashFunction, useBinarySearch);
shardInterval = FastShardPruning(relationId, distributionValueDatum);
}
else
{
@ -940,3 +919,40 @@ TupleToWorkerNode(TupleDesc tupleDescriptor, HeapTuple heapTuple)
return workerNode;
}
/*
* StringToDatum transforms a string representation into a Datum.
*/
Datum
StringToDatum(char *inputString, Oid dataType)
{
Oid typIoFunc = InvalidOid;
Oid typIoParam = InvalidOid;
int32 typeModifier = -1;
Datum datum = 0;
getTypeInputInfo(dataType, &typIoFunc, &typIoParam);
getBaseTypeAndTypmod(dataType, &typeModifier);
datum = OidInputFunctionCall(typIoFunc, inputString, typIoParam, typeModifier);
return datum;
}
/*
* DatumToString returns the string representation of the given datum.
*/
char *
DatumToString(Datum datum, Oid dataType)
{
char *outputString = NULL;
Oid typIoFunc = InvalidOid;
bool typIsVarlena = false;
getTypeOutputInfo(dataType, &typIoFunc, &typIsVarlena);
outputString = OidOutputFunctionCall(typIoFunc, datum);
return outputString;
}

View File

@ -46,4 +46,6 @@ extern void WarnRemoteError(PGconn *connection, PGresult *result);
extern void ReraiseRemoteError(PGconn *connection, PGresult *result);
extern PGconn * ConnectToNode(char *nodeName, int nodePort, char *nodeUser);
extern char * ConnectionGetOptionValue(PGconn *connection, char *optionKeyword);
#endif /* CONNECTION_CACHE_H */

View File

@ -98,5 +98,9 @@ extern void EnsureTablePermissions(Oid relationId, AclMode mode);
extern void EnsureTableOwner(Oid relationId);
extern void EnsureSuperUser(void);
extern bool TableReferenced(Oid relationId);
extern char * ConstructQualifiedShardName(ShardInterval *shardInterval);
extern Datum StringToDatum(char *inputString, Oid dataType);
extern char * DatumToString(Datum datum, Oid dataType);
#endif /* MASTER_METADATA_UTILITY_H */

View File

@ -20,6 +20,7 @@
#include "distributed/shardinterval_utils.h"
#include "nodes/pg_list.h"
#include "distributed/master_metadata_utility.h"
/*
@ -118,6 +119,7 @@ extern Oid ForeignConstraintGetReferencedTableId(char *queryString);
extern void CheckHashPartitionedTable(Oid distributedTableId);
extern void CheckTableSchemaNameForDrop(Oid relationId, char **schemaName,
char **tableName);
extern text * IntegerToText(int32 value);
/* Function declarations for generating metadata for shard and placement creation */
extern Datum master_get_table_metadata(PG_FUNCTION_ARGS);
@ -139,6 +141,7 @@ extern Datum master_drop_all_shards(PG_FUNCTION_ARGS);
/* function declarations for shard creation functionality */
extern Datum master_create_worker_shards(PG_FUNCTION_ARGS);
extern Datum isolate_tenant_to_new_shard(PG_FUNCTION_ARGS);
/* function declarations for shard repair functionality */
extern Datum master_copy_shard_placement(PG_FUNCTION_ARGS);

View File

@ -28,6 +28,7 @@ extern char * DistributionDeleteCommand(char *schemaName,
extern char * TableOwnerResetCommand(Oid distributedRelationId);
extern char * NodeListInsertCommand(List *workerNodeList);
extern List * ShardListInsertCommand(List *shardIntervalList);
extern List * ShardDeleteCommandList(ShardInterval *shardInterval);
extern char * NodeDeleteCommand(uint32 nodeId);
extern char * ColocationIdUpdateCommand(Oid relationId, uint32 colocationId);
extern char * CreateSchemaDDLCommand(Oid schemaId);

View File

@ -41,6 +41,7 @@ extern RangeTblEntry * ExtractSelectRangeTableEntry(Query *query);
extern RangeTblEntry * ExtractInsertRangeTableEntry(Query *query);
extern void AddShardIntervalRestrictionToSelect(Query *subqery,
ShardInterval *shardInterval);
extern ShardInterval * FastShardPruning(Oid distributedTableId, Datum partitionValue);
#endif /* MULTI_ROUTER_PLANNER_H */

View File

@ -155,5 +155,8 @@ extern Datum worker_append_table_to_shard(PG_FUNCTION_ARGS);
extern Datum worker_foreign_file_path(PG_FUNCTION_ARGS);
extern Datum worker_find_block_local_path(PG_FUNCTION_ARGS);
/* Function declaration for calculating hashed value */
extern Datum worker_hash(PG_FUNCTION_ARGS);
#endif /* WORKER_PROTOCOL_H */

View File

@ -366,13 +366,13 @@ SELECT create_distributed_table('get_shardid_test_table2', 'column1');
(1 row)
\COPY get_shardid_test_table2 FROM STDIN with delimiter '|';
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', '{a, b, c}'::text[]);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', '{a, b, c}');
get_shard_id_for_distribution_column
--------------------------------------
540013
(1 row)
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', '{d, e, f}'::text[]);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', '{d, e, f}');
get_shard_id_for_distribution_column
--------------------------------------
540011
@ -394,9 +394,9 @@ SELECT * FROM get_shardid_test_table2_540011;
\c - - - :master_port
-- test mismatching data type
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', 'a'::text);
ERROR: invalid distribution value type
DETAIL: Type of the value does not match the type of the distribution column. Expected type id: 1009, given type id: 25
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', 'a');
ERROR: malformed array literal: "a"
DETAIL: Array value must start with "{" or dimension information.
-- test NULL distribution column value for hash distributed table
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2');
ERROR: distribution value cannot be NULL for tables other than reference tables.
@ -443,13 +443,13 @@ SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', 1);
540014
(1 row)
SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', 'a'::text);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', 'a');
get_shard_id_for_distribution_column
--------------------------------------
540014
(1 row)
SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', '{a, b, c}'::text[]);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', '{a, b, c}');
get_shard_id_for_distribution_column
--------------------------------------
540014

View File

@ -74,6 +74,7 @@ ALTER EXTENSION citus UPDATE TO '6.1-13';
ALTER EXTENSION citus UPDATE TO '6.1-14';
ALTER EXTENSION citus UPDATE TO '6.1-15';
ALTER EXTENSION citus UPDATE TO '6.1-16';
ALTER EXTENSION citus UPDATE TO '6.1-17';
-- ensure no objects were created outside pg_catalog
SELECT COUNT(*)
FROM pg_depend AS pgd,

View File

@ -319,6 +319,29 @@ VACUUM dustbunnies;
WARNING: not propagating VACUUM command to worker nodes
HINT: Set citus.enable_ddl_propagation to true in order to send targeted VACUUM commands to worker nodes.
SET citus.enable_ddl_propagation to DEFAULT;
-- test worker_hash
SELECT worker_hash(123);
worker_hash
-------------
-205084363
(1 row)
SELECT worker_hash('1997-08-08'::date);
worker_hash
-------------
-499701663
(1 row)
-- test a custom type (this test should run after multi_data_types)
SELECT worker_hash('(1, 2)');
ERROR: cannot find a hash function for the input type
HINT: Cast input to a data type with a hash function.
SELECT worker_hash('(1, 2)'::test_composite_type);
worker_hash
-------------
-1895345704
(1 row)
-- TODO: support VERBOSE
-- VACUUM VERBOSE dustbunnies;
-- VACUUM (FULL, VERBOSE) dustbunnies;

View File

@ -135,9 +135,9 @@ test: multi_repair_shards
test: multi_modifications
test: multi_upsert
test: multi_simple_queries
test: multi_utilities
test: multi_create_insert_proxy
test: multi_data_types
test: multi_utilities
test: multi_repartition_udt
test: multi_repartitioned_subquery_udf
test: multi_modifying_xacts

View File

@ -90,9 +90,9 @@ test: multi_repair_shards
test: multi_modifications
test: multi_upsert
test: multi_simple_queries
test: multi_utilities
test: multi_create_insert_proxy
test: multi_data_types
test: multi_utilities
# ---------
# multi_copy creates hash and range-partitioned tables and performs COPY

View File

@ -234,8 +234,8 @@ SELECT create_distributed_table('get_shardid_test_table2', 'column1');
{a, b, c}|1
{d, e, f}|2
\.
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', '{a, b, c}'::text[]);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', '{d, e, f}'::text[]);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', '{a, b, c}');
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', '{d, e, f}');
-- verify result of the get_shard_id_for_distribution_column
\c - - - :worker_1_port
@ -244,7 +244,7 @@ SELECT * FROM get_shardid_test_table2_540011;
\c - - - :master_port
-- test mismatching data type
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', 'a'::text);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2', 'a');
-- test NULL distribution column value for hash distributed table
SELECT get_shard_id_for_distribution_column('get_shardid_test_table2');
@ -268,8 +268,8 @@ SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', NULL);
-- test different data types for reference table
SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', 1);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', 'a'::text);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', '{a, b, c}'::text[]);
SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', 'a');
SELECT get_shard_id_for_distribution_column('get_shardid_test_table4', '{a, b, c}');
-- test range distributed table
CREATE TABLE get_shardid_test_table5(column1 int, column2 int);

View File

@ -74,6 +74,7 @@ ALTER EXTENSION citus UPDATE TO '6.1-13';
ALTER EXTENSION citus UPDATE TO '6.1-14';
ALTER EXTENSION citus UPDATE TO '6.1-15';
ALTER EXTENSION citus UPDATE TO '6.1-16';
ALTER EXTENSION citus UPDATE TO '6.1-17';
-- ensure no objects were created outside pg_catalog
SELECT COUNT(*)

View File

@ -209,6 +209,14 @@ SET citus.enable_ddl_propagation to false;
VACUUM dustbunnies;
SET citus.enable_ddl_propagation to DEFAULT;
-- test worker_hash
SELECT worker_hash(123);
SELECT worker_hash('1997-08-08'::date);
-- test a custom type (this test should run after multi_data_types)
SELECT worker_hash('(1, 2)');
SELECT worker_hash('(1, 2)'::test_composite_type);
-- TODO: support VERBOSE
-- VACUUM VERBOSE dustbunnies;
-- VACUUM (FULL, VERBOSE) dustbunnies;