Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
11 changes: 9 additions & 2 deletions CHANGELOG.rst
Original file line number Diff line number Diff line change
@@ -1,7 +1,14 @@
NVIDIA Model Optimizer Changelog
================================

0.43 (2026-03-xx)
0.44 (2026-05-xx)
^^^^^^^^^^^^^^^^^

**New Features**

- Support full Transformer Engine spec for Minitron pruning (``mcore_minitron``). Now we no longer need to use custom ModelOpt spec. Note that this does not affect the usage of the pruning workflow but makes pruning slightly faster and may result in slightly different pruned model because of different kernel and numerics.

0.43 (2026-04-09)
^^^^^^^^^^^^^^^^^

**Bug Fixes**
Expand Down Expand Up @@ -29,7 +36,7 @@ NVIDIA Model Optimizer Changelog

- Migrated project metadata from ``setup.py`` to a fully declarative ``pyproject.toml``.

0.42 (2026-02-xx)
0.42 (2026-03-10)
^^^^^^^^^^^^^^^^^

**Bug Fixes**
Expand Down
6 changes: 3 additions & 3 deletions examples/megatron_bridge/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,17 +16,17 @@ This directory contains examples of using Model Optimizer with [NeMo Megatron-Br

## Pre-Requisites

Running these examples requires many additional dependencies to be installed (e.g., Megatron-Bridge, Megatron-core, etc.), hence we strongly recommend directly using the NeMo container (e.g., `nvcr.io/nvidia/nemo:26.02`) which has all the dependencies installed.
Running these examples requires many additional dependencies to be installed (e.g., Megatron-Bridge, Megatron-core, etc.), hence we strongly recommend directly using the NeMo container (e.g., `nvcr.io/nvidia/nemo:26.02.01`) which has all the dependencies installed.

To get the latest ModelOpt features and examples scripts, mount your Model-Optimizer repo to the container.
To get the ModelOpt examples scripts, mount your Model-Optimizer repo to the container as follows:

```bash
export MODELOPT_DIR=${PWD}/Model-Optimizer # or set to your local Model-Optimizer repository path if you have cloned it
if [ ! -d "${MODELOPT_DIR}" ]; then
git clone https://github.com/NVIDIA/Model-Optimizer.git ${MODELOPT_DIR}
fi

export DOCKER_IMAGE=nvcr.io/nvidia/nemo:26.02
export DOCKER_IMAGE=nvcr.io/nvidia/nemo:26.02.01
docker run \
--gpus all \
--shm-size=16GB \
Expand Down
4 changes: 2 additions & 2 deletions examples/megatron_bridge/prune_minitron.py
Original file line number Diff line number Diff line change
Expand Up @@ -241,7 +241,7 @@ def main(args: argparse.Namespace):
},
init_model_parallel=True,
)
print_rank_0(f"\nPruning {unwrapped_model=}")
print_rank_0(f"\nPruning model (showing PP rank0): {unwrapped_model}")
print_rank_0(
f"Original model params: {num2hrb(mtp.mcore_minitron.get_mcore_param_count(unwrapped_model))}"
)
Expand Down Expand Up @@ -317,7 +317,7 @@ def score_func_mmlu(m):
else "hybrid_layer_pattern"
)
setattr(provider, hybrid_key, getattr(unwrapped_model, hybrid_key))
print_rank_0(f"\nPruned {unwrapped_model=}")
print_rank_0(f"\nPruned model (showing PP rank0): {unwrapped_model}")
print_rank_0(
f"Pruned model params: {num2hrb(mtp.mcore_minitron.get_mcore_param_count(unwrapped_model))}"
)
Expand Down
Loading
Loading