From 6d406a108dea26c2a755e86fcd09ca048d94d2af Mon Sep 17 00:00:00 2001 From: Daniel Dale Date: Fri, 10 Jun 2022 15:25:59 -0700 Subject: [PATCH 1/4] fix sphinx language warning --- docs/source/conf.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/source/conf.py b/docs/source/conf.py index 8e90952d1..efadb440b 100644 --- a/docs/source/conf.py +++ b/docs/source/conf.py @@ -99,7 +99,7 @@ # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. -language = None +language = "en" # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. From e33a28c3c2ca38f0058060351d3014c2a3128e09 Mon Sep 17 00:00:00 2001 From: Jirka Date: Mon, 13 Jun 2022 15:30:08 +0200 Subject: [PATCH 2/4] fix stale ffmpeg ref #178 --- .azure/ipynb-publish.yml | 1 + .azure/ipynb-tests.yml | 1 + .github/workflows/ci_testing.yml | 1 + 3 files changed, 3 insertions(+) diff --git a/.azure/ipynb-publish.yml b/.azure/ipynb-publish.yml index d21b25f0d..edbb93047 100644 --- a/.azure/ipynb-publish.yml +++ b/.azure/ipynb-publish.yml @@ -62,6 +62,7 @@ jobs: - bash: | set -e + sudo apt-get update -q --fix-missing sudo apt install -y tree ffmpeg pip --version pip install --requirement requirements.txt diff --git a/.azure/ipynb-tests.yml b/.azure/ipynb-tests.yml index 4b0b45b65..091525456 100644 --- a/.azure/ipynb-tests.yml +++ b/.azure/ipynb-tests.yml @@ -42,6 +42,7 @@ jobs: - bash: | set -e + sudo apt-get update -q --fix-missing sudo apt install -y tree ffmpeg pip --version pip install --requirement requirements.txt diff --git a/.github/workflows/ci_testing.yml b/.github/workflows/ci_testing.yml index 5f116f1eb..44ddfd8ac 100644 --- a/.github/workflows/ci_testing.yml +++ b/.github/workflows/ci_testing.yml @@ -39,6 +39,7 @@ jobs: - name: Install dependencies run: | + sudo apt-get update -q --fix-missing sudo apt install -y ffmpeg pip --version pip install --requirement requirements.txt --find-links https://download.pytorch.org/whl/cpu/torch_stable.html From d52e25dfa0cdb8e2e90be3927aaa22c6cff79761 Mon Sep 17 00:00:00 2001 From: Jirka Date: Mon, 13 Jun 2022 15:59:57 +0200 Subject: [PATCH 3/4] link --- lightning_examples/finetuning-scheduler/finetuning-scheduler.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/lightning_examples/finetuning-scheduler/finetuning-scheduler.py b/lightning_examples/finetuning-scheduler/finetuning-scheduler.py index e276fe677..0f03a4f69 100644 --- a/lightning_examples/finetuning-scheduler/finetuning-scheduler.py +++ b/lightning_examples/finetuning-scheduler/finetuning-scheduler.py @@ -59,7 +59,7 @@ # %% [markdown] # ## The Default Finetuning Schedule # -# Schedule definition is facilitated via the [gen_ft_schedule](https://finetuning-scheduler.readthedocs.io/en/stable/api/finetuning_scheduler.fts_supporters.html#finetuning_scheduler.fts_supporters.SchedulingMixin.gen_ft_schedule) method which dumps a default finetuning schedule (by default using a naive, 2-parameters per level heuristic) which can be adjusted as +# Schedule definition is facilitated via the [gen_ft_schedule](https://finetuning-scheduler.readthedocs.io/en/stable/api/finetuning_scheduler.fts_supporters.html#finetuning_scheduler.fts_supporters.ScheduleImplMixin.gen_ft_schedule) method which dumps a default finetuning schedule (by default using a naive, 2-parameters per level heuristic) which can be adjusted as # desired by the user and/or subsequently passed to the callback. Using the default/implicitly generated schedule will likely be less computationally efficient than a user-defined finetuning schedule but is useful for exploring a model's finetuning behavior and can serve as a good baseline for subsequent explicit schedule refinement. # While the current version of [FinetuningScheduler](https://finetuning-scheduler.readthedocs.io/en/stable/api/finetuning_scheduler.fts.html#finetuning_scheduler.fts.FinetuningScheduler) only supports single optimizer and (optional) lr_scheduler configurations, per-phase maximum learning rates can be set as demonstrated in the next section. From 36aef23f7f15022d2e9d21e3a98e110e968ba331 Mon Sep 17 00:00:00 2001 From: Jirka Date: Mon, 13 Jun 2022 16:25:46 +0200 Subject: [PATCH 4/4] hydra --- lightning_examples/finetuning-scheduler/.meta.yml | 1 + 1 file changed, 1 insertion(+) diff --git a/lightning_examples/finetuning-scheduler/.meta.yml b/lightning_examples/finetuning-scheduler/.meta.yml index 3cae7999e..041e86581 100644 --- a/lightning_examples/finetuning-scheduler/.meta.yml +++ b/lightning_examples/finetuning-scheduler/.meta.yml @@ -15,5 +15,6 @@ description: | and foundational model weights. The required dependencies are installed via the finetuning-scheduler ``[examples]`` extra. requirements: - finetuning-scheduler[examples] + - hydra-core>=1.1.0 accelerator: - GPU