Merge pull request #3344 from citusdata/fix-extension-already-exists-test

Fix tests when hll/topn installed
pull/3307/head
Philip Dubé 2019-12-24 21:45:36 +00:00 committed by GitHub
commit 11368451f4
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
7 changed files with 62 additions and 159 deletions

View File

@ -10,13 +10,6 @@ WHERE name = 'hll'
\gset
:create_cmd;
ERROR: extension "hll" already exists
\c - - - :worker_1_port
:create_cmd;
ERROR: extension "hll" already exists
\c - - - :worker_2_port
:create_cmd;
ERROR: extension "hll" already exists
\c - - - :master_port
SET citus.shard_count TO 4;
CREATE TABLE raw_table (day date, user_id int);
CREATE TABLE daily_uniques(day date, unique_users hll);
@ -126,22 +119,22 @@ GROUP BY(1);
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360289 daily_uniques
-> Seq Scan on daily_uniques_360615 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360290 daily_uniques
-> Seq Scan on daily_uniques_360616 daily_uniques
-> Task
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360291 daily_uniques
-> Seq Scan on daily_uniques_360617 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360292 daily_uniques
-> Seq Scan on daily_uniques_360618 daily_uniques
(25 rows)
SET hll.force_groupagg to ON;
@ -164,22 +157,22 @@ GROUP BY(1);
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360289 daily_uniques
-> Seq Scan on daily_uniques_360615 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360290 daily_uniques
-> Seq Scan on daily_uniques_360616 daily_uniques
-> Task
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360291 daily_uniques
-> Seq Scan on daily_uniques_360617 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360292 daily_uniques
-> Seq Scan on daily_uniques_360618 daily_uniques
(27 rows)
-- Test disabling hash_agg with operator on coordinator query
@ -201,22 +194,22 @@ GROUP BY(1);
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360289 daily_uniques
-> Seq Scan on daily_uniques_360615 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360290 daily_uniques
-> Seq Scan on daily_uniques_360616 daily_uniques
-> Task
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360291 daily_uniques
-> Seq Scan on daily_uniques_360617 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360292 daily_uniques
-> Seq Scan on daily_uniques_360618 daily_uniques
(25 rows)
SET hll.force_groupagg to ON;
@ -239,22 +232,22 @@ GROUP BY(1);
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360289 daily_uniques
-> Seq Scan on daily_uniques_360615 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360290 daily_uniques
-> Seq Scan on daily_uniques_360616 daily_uniques
-> Task
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360291 daily_uniques
-> Seq Scan on daily_uniques_360617 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360292 daily_uniques
-> Seq Scan on daily_uniques_360618 daily_uniques
(27 rows)
-- Test disabling hash_agg with expression on coordinator query
@ -276,22 +269,22 @@ GROUP BY(1);
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360289 daily_uniques
-> Seq Scan on daily_uniques_360615 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360290 daily_uniques
-> Seq Scan on daily_uniques_360616 daily_uniques
-> Task
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360291 daily_uniques
-> Seq Scan on daily_uniques_360617 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360292 daily_uniques
-> Seq Scan on daily_uniques_360618 daily_uniques
(25 rows)
SET hll.force_groupagg to ON;
@ -314,22 +307,22 @@ GROUP BY(1);
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360289 daily_uniques
-> Seq Scan on daily_uniques_360615 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360290 daily_uniques
-> Seq Scan on daily_uniques_360616 daily_uniques
-> Task
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360291 daily_uniques
-> Seq Scan on daily_uniques_360617 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360292 daily_uniques
-> Seq Scan on daily_uniques_360618 daily_uniques
(27 rows)
-- Test disabling hash_agg with having
@ -351,22 +344,22 @@ GROUP BY(1);
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360289 daily_uniques
-> Seq Scan on daily_uniques_360615 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360290 daily_uniques
-> Seq Scan on daily_uniques_360616 daily_uniques
-> Task
Node: host=localhost port=57637 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360291 daily_uniques
-> Seq Scan on daily_uniques_360617 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> HashAggregate
Group Key: day
-> Seq Scan on daily_uniques_360292 daily_uniques
-> Seq Scan on daily_uniques_360618 daily_uniques
(25 rows)
SET hll.force_groupagg to ON;
@ -394,7 +387,7 @@ HAVING hll_cardinality(hll_union_agg(unique_users)) > 1;
Filter: (hll_cardinality(hll_union_agg(unique_users)) > '1'::double precision)
-> Sort
Sort Key: day
-> Seq Scan on daily_uniques_360289 daily_uniques
-> Seq Scan on daily_uniques_360615 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> GroupAggregate
@ -402,7 +395,7 @@ HAVING hll_cardinality(hll_union_agg(unique_users)) > 1;
Filter: (hll_cardinality(hll_union_agg(unique_users)) > '1'::double precision)
-> Sort
Sort Key: day
-> Seq Scan on daily_uniques_360290 daily_uniques
-> Seq Scan on daily_uniques_360616 daily_uniques
-> Task
Node: host=localhost port=57637 dbname=regression
-> GroupAggregate
@ -410,7 +403,7 @@ HAVING hll_cardinality(hll_union_agg(unique_users)) > 1;
Filter: (hll_cardinality(hll_union_agg(unique_users)) > '1'::double precision)
-> Sort
Sort Key: day
-> Seq Scan on daily_uniques_360291 daily_uniques
-> Seq Scan on daily_uniques_360617 daily_uniques
-> Task
Node: host=localhost port=57638 dbname=regression
-> GroupAggregate
@ -418,7 +411,7 @@ HAVING hll_cardinality(hll_union_agg(unique_users)) > 1;
Filter: (hll_cardinality(hll_union_agg(unique_users)) > '1'::double precision)
-> Sort
Sort Key: day
-> Seq Scan on daily_uniques_360292 daily_uniques
-> Seq Scan on daily_uniques_360618 daily_uniques
(40 rows)
DROP TABLE raw_table;
@ -432,11 +425,6 @@ AS create_topn FROM pg_available_extensions()
WHERE name = 'topn'
\gset
:create_topn;
\c - - - :worker_1_port
:create_topn;
\c - - - :worker_2_port
:create_topn;
\c - - - :master_port
CREATE TABLE customer_reviews (day date, user_id int, review int);
CREATE TABLE popular_reviewer(day date, reviewers jsonb);
SELECT create_distributed_table('customer_reviews', 'user_id');

View File

@ -14,21 +14,6 @@ WHERE name = 'hll'
f
(1 row)
\c - - - :worker_1_port
:create_cmd;
hll_present
-------------
f
(1 row)
\c - - - :worker_2_port
:create_cmd;
hll_present
-------------
f
(1 row)
\c - - - :master_port
SET citus.shard_count TO 4;
CREATE TABLE raw_table (day date, user_id int);
CREATE TABLE daily_uniques(day date, unique_users hll);
@ -210,21 +195,6 @@ WHERE name = 'topn'
f
(1 row)
\c - - - :worker_1_port
:create_topn;
topn_present
--------------
f
(1 row)
\c - - - :worker_2_port
:create_topn;
topn_present
--------------
f
(1 row)
\c - - - :master_port
CREATE TABLE customer_reviews (day date, user_id int, review int);
CREATE TABLE popular_reviewer(day date, reviewers jsonb);
SELECT create_distributed_table('customer_reviews', 'user_id');

View File

@ -10,13 +10,6 @@ WHERE name = 'hll'
\gset
:create_cmd;
ERROR: extension "hll" already exists
\c - - - :worker_1_port
:create_cmd;
ERROR: extension "hll" already exists
\c - - - :worker_2_port
:create_cmd;
ERROR: extension "hll" already exists
\c - - - :master_port
SET citus.shard_count TO 4;
CREATE TABLE raw_table (day date, user_id int);
CREATE TABLE daily_uniques(day date, unique_users hll);
@ -432,11 +425,6 @@ AS create_topn FROM pg_available_extensions()
WHERE name = 'topn'
\gset
:create_topn;
\c - - - :worker_1_port
:create_topn;
\c - - - :worker_2_port
:create_topn;
\c - - - :master_port
CREATE TABLE customer_reviews (day date, user_id int, review int);
CREATE TABLE popular_reviewer(day date, reviewers jsonb);
SELECT create_distributed_table('customer_reviews', 'user_id');

View File

@ -9,11 +9,6 @@ AS create_cmd FROM pg_available_extensions()
WHERE name = 'hll'
\gset
:create_cmd;
\c - - - :worker_1_port
:create_cmd;
\c - - - :worker_2_port
:create_cmd;
\c - - - :master_port
-- Try to execute count(distinct) when approximate distincts aren't enabled
SELECT count(distinct l_orderkey) FROM lineitem;
count

View File

@ -14,21 +14,6 @@ WHERE name = 'hll'
f
(1 row)
\c - - - :worker_1_port
:create_cmd;
hll_present
-------------
f
(1 row)
\c - - - :worker_2_port
:create_cmd;
hll_present
-------------
f
(1 row)
\c - - - :master_port
-- Try to execute count(distinct) when approximate distincts aren't enabled
SELECT count(distinct l_orderkey) FROM lineitem;
count

View File

@ -11,14 +11,6 @@ WHERE name = 'hll'
:create_cmd;
\c - - - :worker_1_port
:create_cmd;
\c - - - :worker_2_port
:create_cmd;
\c - - - :master_port
SET citus.shard_count TO 4;
CREATE TABLE raw_table (day date, user_id int);
@ -160,13 +152,6 @@ WHERE name = 'topn'
:create_topn;
\c - - - :worker_1_port
:create_topn;
\c - - - :worker_2_port
:create_topn;
\c - - - :master_port
CREATE TABLE customer_reviews (day date, user_id int, review int);
CREATE TABLE popular_reviewer(day date, reviewers jsonb);

View File

@ -13,14 +13,6 @@ WHERE name = 'hll'
:create_cmd;
\c - - - :worker_1_port
:create_cmd;
\c - - - :worker_2_port
:create_cmd;
\c - - - :master_port
-- Try to execute count(distinct) when approximate distincts aren't enabled
SELECT count(distinct l_orderkey) FROM lineitem;