fix(stdlib): fix issues when using LVGL TLSF memory pool to manage more than 4 GiB of memory #5720
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description of the feature or fix
People will know that I'm developing UEFI applications with LVGL in recent days after read #5391 (comment).
Because UEFI memory pool will cause many memory fragments. So, my project uses LVGL builtin TLSF memory pool to manage the memory. And I found if I use LVGL TLSF memory pool to manage more than 4 GiB of memory, I cannot get the correct result via lv_mem_monitor API.
I also found the original LVGL implementation have considered about the 64-bit environments and many lv_mem APIs use size_t to manage size. So, I think it should not break the ABI for 32-bit environments if we use size_t type instead of uint32_t type to fix the issue. (For 64-bit LVGL users needs to recompile. But I think most people don't use 64-bit LVGL instance in the current stage. It should be acceptable.)
Notes
lv_conf_template.h
run lv_conf_internal_gen.py and update Kconfig.scripts/code-format.py
(astyle version v3.4.10 needs to be installed) and follow the Code Conventions.Kenji Mouri