Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Resolve issues #5 and #1: reduce number of collisions in the ptrack map #6

Merged
merged 7 commits into from
May 16, 2021

Conversation

ololobus
Copy link
Contributor

@ololobus ololobus commented Apr 22, 2021

Resolve #5
Resolve #1

…slots.

Previously we thought that 1 MB can track changes page-to-page in the 1 GB of
data files. However, recently it became evident that our ptrack map or basic
hash table behaves more like a Bloom filter with a number of hash functions k = 1.
See more here: https://en.wikipedia.org/wiki/Bloom_filter#Probability_of_false_positives.
Such filter has naturally more collisions.

By storing update_lsn of each block in the additional slot we perform as
a Bloom filter with k = 2, which significatly reduces collision rate.
@ololobus ololobus force-pushed the double_slot branch 3 times, most recently from 4c32e9e to 38aa439 Compare April 22, 2021 21:33
@ololobus ololobus changed the title Resolve issue#5: reduce number of collisions in the ptrack map Resolve issue #5 and #1: reduce number of collisions in the ptrack map Apr 22, 2021
@ololobus ololobus changed the title Resolve issue #5 and #1: reduce number of collisions in the ptrack map Resolve issues #5 and #1: reduce number of collisions in the ptrack map Apr 22, 2021
ptrack.c Outdated
Comment on lines 537 to 538
update_lsn1 = pg_atomic_read_u64(&ptrack_map->entries[slot1]);
update_lsn2 = pg_atomic_read_u64(&ptrack_map->entries[slot2]);
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It is better to fetch and check slot1 first, and only if check passed then fetch and check slot2.
This way you will save TLB and cache misses for slot2 for most of page items.
Note that compiler could not optimize/reorder atomic instructions.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

OK, I hope that I did it

FROM
(SELECT count(path) AS changed_files,
sum(
length(replace(right((pagemap)::text, -1)::varbit::text, '0', ''))
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Если таблицы 8TB, то вот эта строчка потребует выделение 1GB памяти для преобразования ::varbit::text.
Соответственно, таблица 16TB потребует уже 2GB памяти, и постгресс просто сам не позволит этого сделать.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Это очень грустно, что varbit не имеет функции countbits.

Copy link
Contributor

@funny-falcon funny-falcon May 13, 2021

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

В любом случае, для ptrack_get_change_stat и ptrack_get_change_file_stat кажется нужно создать ptrack_get_pagecount (ну или другое название).
Или даже просто реализовать ptrack_get_change_file_stat полностью в сишке.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Таблицы же разбиты на сегменты по 1 ГБ дефолтно, а ptrack_get_pagemapset() выдаёт изначально битмапы per file/segment, то есть потребуется максимум в 1000 раз меньше памяти на каждое преобразование. Разве нет?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

А ок. Я ещё не посмотрел ptrack_get_pagemapset() .

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Слушай, но я бы всё равно поменял бы ptrack_get_pagemapset, добавив поле count в вывод.
pg_probackup при этом не поломается, т.к. он указывает поля, которые хочет.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Сделал


/* Delete and try again */
durable_unlink(ptrack_path, LOG);
is_new_map = true;
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Не могу найти, где делается unmap в этом случае?
При этом сразу после метки ptrack_map_reinit делается durable_unlink(ptrack_mmap_path).
В итоге, этот файл повисает невидимкой в файловой системе, и в адрессном пространстве процесса повисает его mmap.

Наверное есть смысл позвать здесь ptrackCleanFilesAndMap ?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Да, похоже на то. Я сомневался в этом месте, но потом забыл и не разобрался до конца

@ololobus
Copy link
Contributor Author

Everything seems to be working, so I'm merging this one. If the internal QA finds out anything, we will fix it in master or with another PR

@ololobus ololobus merged commit 708c8e2 into master May 16, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Reduce number of collisions in the ptrack map Human-readable changeset
2 participants