Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bump postgres versions used in CI to latest #3491

Merged
merged 1 commit into from Aug 17, 2021
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
2 changes: 1 addition & 1 deletion .github/workflows/coverity.yaml
Expand Up @@ -13,7 +13,7 @@ jobs:
strategy:
fail-fast: false
matrix:
pg: ["12.7", "13.3"]
pg: ["12.8", "13.4"]
os: [ubuntu-20.04]
env:
PG_SRC_DIR: pgbuild
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/cron-tests.yaml
Expand Up @@ -37,7 +37,7 @@ jobs:
strategy:
fail-fast: false
env:
PG_VERSION: 12.6
PG_VERSION: 12.8

steps:
- name: Checkout TimescaleDB
Expand Down
6 changes: 3 additions & 3 deletions .github/workflows/linux-32bit-build-and-test.yaml
Expand Up @@ -19,13 +19,13 @@ jobs:
strategy:
fail-fast: false
matrix:
pg: [ "12.7", "13.3" ]
pg: [ "12.8", "13.4" ]
build_type: [ Debug ]
include:
- pg: 12.7
- pg: 12.8
ignores: append-12 chunk_adaptive continuous_aggs_bgw_drop_chunks remote_txn transparent_decompression-12 plan_skip_scan-12
pg_major: 12
- pg: 13.3
- pg: 13.4
ignores: append-13 chunk_adaptive remote_txn transparent_decompression-13 vacuum_parallel plan_skip_scan-13
pg_major: 13

Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/sanitizer-build-and-test.yaml
Expand Up @@ -43,7 +43,7 @@ jobs:
matrix:
# "os" has to be in the matrix due to a bug in "env": https://github.community/t/how-to-use-env-context/16975
os: ["ubuntu-20.04"]
pg: ["12.7", "13.3"]
pg: ["12.8", "13.4"]
steps:
- name: Install Linux Dependencies
run: |
Expand Down
12 changes: 6 additions & 6 deletions .github/workflows/update-test.yaml
Expand Up @@ -13,11 +13,11 @@ jobs:
runs-on: 'ubuntu-latest'
strategy:
matrix:
pg: ["12.7","13.3"]
pg: ["12.8","13.4"]
include:
- pg: 12.7
- pg: 12.8
pg_major: 12
- pg: 13.3
- pg: 13.4
pg_major: 13
fail-fast: false
env:
Expand Down Expand Up @@ -48,11 +48,11 @@ jobs:
runs-on: 'ubuntu-latest'
strategy:
matrix:
pg: ["12.7","13.3"]
pg: ["12.8","13.4"]
include:
- pg: 12.7
- pg: 12.8
pg_major: 12
- pg: 13.3
- pg: 13.4
pg_major: 13
fail-fast: false
env:
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/windows-build-and-test.yaml
Expand Up @@ -17,9 +17,9 @@ jobs:
build_type: [ Debug, Release ]
include:
- pg: 12
pkg_version: 12.7.1
pkg_version: 12.8.1
- pg: 13
pkg_version: 13.3.1
pkg_version: 13.4.1
env:
# PostgreSQL configuration
PGPORT: 6543
Expand Down
12 changes: 8 additions & 4 deletions scripts/gh_matrix_builder.py
Expand Up @@ -21,9 +21,9 @@
event_type = sys.argv[1]

PG12_EARLIEST = "12.0"
PG12_LATEST = "12.7"
PG12_LATEST = "12.8"
PG13_EARLIEST = "13.2"
PG13_LATEST = "13.3"
PG13_LATEST = "13.4"

m = {"include": [],}

Expand Down Expand Up @@ -118,12 +118,16 @@ def macos_config(overrides):
# add debug test for first supported PG12 version
pg12_debug_earliest = {
"pg": PG12_EARLIEST,
"installcheck_args": "IGNORES='cluster-12'"
"installcheck_args": "IGNORES='cluster-12 compression_ddl continuous_aggs_concurrent_refresh continuous_aggs_insert continuous_aggs_multi deadlock_drop_chunks_compress deadlock_dropchunks_select dist_restore_point dropchunks_race insert_dropchunks_race isolation_nop multi_transaction_indexing read_committed_insert read_uncommitted_insert remote_create_chunk reorder_deadlock reorder_vs_insert reorder_vs_insert_other_chunk reorder_vs_select repeatable_read_insert serializable_insert serializable_insert_rollback'"
}
m["include"].append(build_debug_config(pg12_debug_earliest))

# add debug test for first supported PG13 version
m["include"].append(build_debug_config({"pg":PG13_EARLIEST}))
pg13_debug_earliest = {
"pg": PG13_EARLIEST,
"installcheck_args": "IGNORES='compression_ddl continuous_aggs_concurrent_refresh continuous_aggs_insert continuous_aggs_multi deadlock_drop_chunks_compress deadlock_dropchunks_select dist_restore_point dropchunks_race insert_dropchunks_race isolation_nop multi_transaction_indexing read_committed_insert read_uncommitted_insert remote_create_chunk reorder_deadlock reorder_vs_insert reorder_vs_insert_other_chunk reorder_vs_select repeatable_read_insert serializable_insert serializable_insert_rollback'"
}
m["include"].append(build_debug_config(pg13_debug_earliest))

# add debug test for MacOS
m["include"].append(build_debug_config(macos_config({})))
Expand Down
50 changes: 35 additions & 15 deletions test/isolation/expected/deadlock_dropchunks_select.out
Expand Up @@ -2,32 +2,42 @@ Parsed test spec with 2 sessions

starting permutation: s1a s1b s2a s2b
step s1a: SELECT count (*) FROM drop_chunks('dt', '2018-12-25 00:00'::timestamptz);
count
count
-----
24
(1 row)

24
step s1b: COMMIT;
step s2a: SELECT typ, loc, mtim FROM DT , SL , ST WHERE SL.lid = DT.lid AND ST.sid = DT.sid AND mtim >= '2018-12-01 03:00:00+00' AND mtim <= '2018-12-01 04:00:00+00' AND typ = 'T1' ;
typ loc mtim
typ|loc|mtim
---+---+----
(0 rows)

step s2b: COMMIT;

starting permutation: s1a s2a s1b s2b
step s1a: SELECT count (*) FROM drop_chunks('dt', '2018-12-25 00:00'::timestamptz);
count
count
-----
24
(1 row)

24
step s2a: SELECT typ, loc, mtim FROM DT , SL , ST WHERE SL.lid = DT.lid AND ST.sid = DT.sid AND mtim >= '2018-12-01 03:00:00+00' AND mtim <= '2018-12-01 04:00:00+00' AND typ = 'T1' ; <waiting ...>
step s1b: COMMIT;
step s2a: <... completed>
typ loc mtim
typ|loc|mtim
---+---+----
(0 rows)

step s2b: COMMIT;

starting permutation: s1a s2a s2b s1b
step s1a: SELECT count (*) FROM drop_chunks('dt', '2018-12-25 00:00'::timestamptz);
count
count
-----
24
(1 row)

24
step s2a: SELECT typ, loc, mtim FROM DT , SL , ST WHERE SL.lid = DT.lid AND ST.sid = DT.sid AND mtim >= '2018-12-01 03:00:00+00' AND mtim <= '2018-12-01 04:00:00+00' AND typ = 'T1' ; <waiting ...>
step s2a: <... completed>
ERROR: canceling statement due to lock timeout
Expand All @@ -36,7 +46,9 @@ step s1b: COMMIT;

starting permutation: s2a s1a s1b s2b
step s2a: SELECT typ, loc, mtim FROM DT , SL , ST WHERE SL.lid = DT.lid AND ST.sid = DT.sid AND mtim >= '2018-12-01 03:00:00+00' AND mtim <= '2018-12-01 04:00:00+00' AND typ = 'T1' ;
typ loc mtim
typ|loc|mtim
---+---+----
(0 rows)

step s1a: SELECT count (*) FROM drop_chunks('dt', '2018-12-25 00:00'::timestamptz); <waiting ...>
step s1a: <... completed>
Expand All @@ -46,23 +58,31 @@ step s2b: COMMIT;

starting permutation: s2a s1a s2b s1b
step s2a: SELECT typ, loc, mtim FROM DT , SL , ST WHERE SL.lid = DT.lid AND ST.sid = DT.sid AND mtim >= '2018-12-01 03:00:00+00' AND mtim <= '2018-12-01 04:00:00+00' AND typ = 'T1' ;
typ loc mtim
typ|loc|mtim
---+---+----
(0 rows)

step s1a: SELECT count (*) FROM drop_chunks('dt', '2018-12-25 00:00'::timestamptz); <waiting ...>
step s2b: COMMIT;
step s1a: <... completed>
count
count
-----
24
(1 row)

24
step s1b: COMMIT;

starting permutation: s2a s2b s1a s1b
step s2a: SELECT typ, loc, mtim FROM DT , SL , ST WHERE SL.lid = DT.lid AND ST.sid = DT.sid AND mtim >= '2018-12-01 03:00:00+00' AND mtim <= '2018-12-01 04:00:00+00' AND typ = 'T1' ;
typ loc mtim
typ|loc|mtim
---+---+----
(0 rows)

step s2b: COMMIT;
step s1a: SELECT count (*) FROM drop_chunks('dt', '2018-12-25 00:00'::timestamptz);
count
count
-----
24
(1 row)

24
step s1b: COMMIT;
100 changes: 69 additions & 31 deletions test/isolation/expected/dropchunks_race.out
Expand Up @@ -3,88 +3,126 @@ Parsed test spec with 5 sessions
starting permutation: s3_chunks_found_wait s1_drop_chunks s2_drop_chunks s3_chunks_found_release s3_show_missing_slices s3_show_num_chunks s3_show_data
step s3_chunks_found_wait: SELECT debug_waitpoint_enable('drop_chunks_chunks_found');
debug_waitpoint_enable
----------------------

(1 row)


step s1_drop_chunks: SELECT count(*) FROM drop_chunks('dropchunks_race_t1', TIMESTAMPTZ '2020-03-01'); <waiting ...>
step s2_drop_chunks: SELECT count(*) FROM drop_chunks('dropchunks_race_t1', TIMESTAMPTZ '2020-03-01'); <waiting ...>
step s3_chunks_found_release: SELECT debug_waitpoint_release('drop_chunks_chunks_found');
debug_waitpoint_release
-----------------------

(1 row)


step s1_drop_chunks: <... completed>
count
count
-----
1
(1 row)

1
step s2_drop_chunks: <... completed>
count
count
-----
0
(1 row)

0
step s3_show_missing_slices: SELECT count(*) FROM _timescaledb_catalog.chunk_constraint WHERE dimension_slice_id NOT IN (SELECT id FROM _timescaledb_catalog.dimension_slice);
count
count
-----
0
(1 row)

0
step s3_show_num_chunks: SELECT count(*) FROM show_chunks('dropchunks_race_t1') ORDER BY 1;
count
count
-----
0
(1 row)

0
step s3_show_data: SELECT * FROM dropchunks_race_t1 ORDER BY 1;
time device temp
time|device|temp
----+------+----
(0 rows)


starting permutation: s4_chunks_dropped_wait s1_drop_chunks s5_insert_new_chunk s4_chunks_dropped_release s3_show_missing_slices s3_show_num_chunks s3_show_data
step s4_chunks_dropped_wait: SELECT debug_waitpoint_enable('drop_chunks_end');
debug_waitpoint_enable
----------------------

(1 row)


step s1_drop_chunks: SELECT count(*) FROM drop_chunks('dropchunks_race_t1', TIMESTAMPTZ '2020-03-01'); <waiting ...>
step s5_insert_new_chunk: INSERT INTO dropchunks_race_t1 VALUES ('2020-03-01 10:30', 1, 2.2); <waiting ...>
step s4_chunks_dropped_release: SELECT debug_waitpoint_release('drop_chunks_end');
debug_waitpoint_release
-----------------------

(1 row)


step s1_drop_chunks: <... completed>
count
count
-----
1
(1 row)

1
step s5_insert_new_chunk: <... completed>
step s3_show_missing_slices: SELECT count(*) FROM _timescaledb_catalog.chunk_constraint WHERE dimension_slice_id NOT IN (SELECT id FROM _timescaledb_catalog.dimension_slice);
count
count
-----
0
(1 row)

0
step s3_show_num_chunks: SELECT count(*) FROM show_chunks('dropchunks_race_t1') ORDER BY 1;
count
count
-----
1
(1 row)

1
step s3_show_data: SELECT * FROM dropchunks_race_t1 ORDER BY 1;
time device temp
time |device|temp
----------------------------+------+----
Sun Mar 01 10:30:00 2020 PST| 1| 2.2
(1 row)

Sun Mar 01 10:30:00 2020 PST1 2.2

starting permutation: s4_chunks_dropped_wait s1_drop_chunks s5_insert_old_chunk s4_chunks_dropped_release s3_show_missing_slices s3_show_num_chunks s3_show_data
step s4_chunks_dropped_wait: SELECT debug_waitpoint_enable('drop_chunks_end');
debug_waitpoint_enable
----------------------

(1 row)


step s1_drop_chunks: SELECT count(*) FROM drop_chunks('dropchunks_race_t1', TIMESTAMPTZ '2020-03-01'); <waiting ...>
step s5_insert_old_chunk: INSERT INTO dropchunks_race_t1 VALUES ('2020-01-02 10:31', 1, 1.1); <waiting ...>
step s4_chunks_dropped_release: SELECT debug_waitpoint_release('drop_chunks_end');
debug_waitpoint_release
-----------------------

(1 row)


step s1_drop_chunks: <... completed>
count
count
-----
1
(1 row)

1
step s5_insert_old_chunk: <... completed>
step s3_show_missing_slices: SELECT count(*) FROM _timescaledb_catalog.chunk_constraint WHERE dimension_slice_id NOT IN (SELECT id FROM _timescaledb_catalog.dimension_slice);
count
count
-----
0
(1 row)

0
step s3_show_num_chunks: SELECT count(*) FROM show_chunks('dropchunks_race_t1') ORDER BY 1;
count
count
-----
1
(1 row)

1
step s3_show_data: SELECT * FROM dropchunks_race_t1 ORDER BY 1;
time device temp
time |device|temp
----------------------------+------+----
Thu Jan 02 10:31:00 2020 PST| 1| 1.1
(1 row)

Thu Jan 02 10:31:00 2020 PST1 1.1
12 changes: 8 additions & 4 deletions test/isolation/expected/insert_dropchunks_race.out
Expand Up @@ -5,11 +5,15 @@ step s1a: INSERT INTO insert_dropchunks_race_t1 VALUES ('2020-01-03 10:30', 3, 3
step s2a: SELECT COUNT(*) FROM drop_chunks('insert_dropchunks_race_t1', TIMESTAMPTZ '2020-03-01'); <waiting ...>
step s1b: COMMIT;
step s2a: <... completed>
count
count
-----
2
(1 row)

2
step s2b: COMMIT;
step s1c: SELECT COUNT(*) FROM _timescaledb_catalog.chunk_constraint LEFT JOIN _timescaledb_catalog.dimension_slice sl ON dimension_slice_id = sl.id WHERE sl.id IS NULL;
count
count
-----
0
(1 row)

0