From cd7495346b844bf1ac50a3b663f5510a723e3462 Mon Sep 17 00:00:00 2001 From: Jack Zhang <32371937+jackzhxng@users.noreply.github.com> Date: Tue, 11 Feb 2025 11:10:24 -0800 Subject: [PATCH 1/2] Fix logging.info not appearing --- examples/models/llama/export_llama.py | 11 ++++++----- 1 file changed, 6 insertions(+), 5 deletions(-) diff --git a/examples/models/llama/export_llama.py b/examples/models/llama/export_llama.py index eeb425c338c..5678960b18a 100644 --- a/examples/models/llama/export_llama.py +++ b/examples/models/llama/export_llama.py @@ -7,6 +7,12 @@ # Example script for exporting Llama2 to flatbuffer import logging + +# force=True to ensure logging while in debugger. Set up logger before any +# other imports. +FORMAT = "[%(levelname)s %(asctime)s %(filename)s:%(lineno)s] %(message)s" +logging.basicConfig(level=logging.INFO, format=FORMAT, force=True) + import sys import torch @@ -15,11 +21,6 @@ sys.setrecursionlimit(4096) - -FORMAT = "[%(levelname)s %(asctime)s %(filename)s:%(lineno)s] %(message)s" -logging.basicConfig(level=logging.INFO, format=FORMAT) - - def main() -> None: seed = 42 torch.manual_seed(seed) From 96140a1dd2d944c7ff50c9e17ab904444c71482b Mon Sep 17 00:00:00 2001 From: Jack Zhang <32371937+jackzhxng@users.noreply.github.com> Date: Tue, 11 Feb 2025 11:40:42 -0800 Subject: [PATCH 2/2] Lint --- examples/models/llama/export_llama.py | 1 + 1 file changed, 1 insertion(+) diff --git a/examples/models/llama/export_llama.py b/examples/models/llama/export_llama.py index 5678960b18a..e25a8a007eb 100644 --- a/examples/models/llama/export_llama.py +++ b/examples/models/llama/export_llama.py @@ -21,6 +21,7 @@ sys.setrecursionlimit(4096) + def main() -> None: seed = 42 torch.manual_seed(seed)