Add convert_llama3_config_to_decilm_config + unit test #465
Add convert_llama3_config_to_decilm_config + unit test #465danielkorzekwa merged 30 commits intofeature/compressfrom
Conversation
using MIP-based NAS search algorithm. Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## feature/compress #465 +/- ##
=================================================
Coverage 73.40% 73.40%
=================================================
Files 180 180
Lines 18127 18127
=================================================
Hits 13306 13306
Misses 4821 4821 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
| return Path(request.config.rootpath) | ||
|
|
||
|
|
||
| def test_convert_llama3_config_to_decilm_config(project_root_path: Path, tmp_path: Path): |
There was a problem hiding this comment.
Can we extend this and take a dummy input (available via model.dummy_inputs) and run forward pass on llama model (output1 = model(**model.dummy_inputs)) and converted DeciLM model and assert they are same / almost same (torch.allclose)? This will be a more robuts check that conversion is correct
There was a problem hiding this comment.
We have a util function to do this. from _test_utils.torch.transformers_models import tf_output_tester
There was a problem hiding this comment.
regarding validating on dummy input (this is planned in the internal nvidia repo, let's migrate it later), creating an issue to migrate it. issues/14
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
…ation. Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
…ress module. Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
…ntal/ folder to not be run by CICD yet. Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Keval Morabia <28916987+kevalmorabia97@users.noreply.github.com>
…tmp_path. Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
…o_decilm_convertion
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
…o_decilm_convertion
…rtion Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
| @pytest.fixture | ||
| def project_root_path(request: pytest.FixtureRequest) -> Path: | ||
| return Path(request.config.rootpath) |
There was a problem hiding this comment.
We can define this in tests/experimental/torch/_compress/conftest.py so it can be reused in both tests
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
Signed-off-by: Daniel Korzekwa <dkorzekwa@nvidia.com>
What does this PR do?
Additional Information
First review/merge #464 and merge it back to this feature branch.