-
Notifications
You must be signed in to change notification settings - Fork 389
fix: layerwise calibration backward-compat, recipe split, batch-size guard #1310
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -955,6 +955,25 @@ def quantize_main( | |
| default_pad_token, | ||
| device: torch.device, | ||
| ): | ||
| # Load the recipe up front so we can detect layerwise calibration before batch-size probing. | ||
| recipe = None | ||
| if args.recipe is not None and not args.auto_quantize_bits: | ||
| print(f"Use recipe {args.recipe} for quantization") | ||
| recipe = load_recipe(args.recipe) | ||
| if not isinstance(recipe, ModelOptPTQRecipe): | ||
| raise TypeError( | ||
| f"Expected PTQ recipe, but got {type(recipe).__name__} from {args.recipe}" | ||
| ) | ||
|
|
||
| def _is_layerwise(obj): | ||
| if isinstance(obj, ModelOptPTQRecipe): | ||
| return _is_layerwise(obj.quantize.algorithm) | ||
| if isinstance(obj, list): | ||
| return any(_is_layerwise(a) for a in obj) | ||
| return bool(getattr(obj, "layerwise", False)) | ||
|
Comment on lines
+968
to
+973
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Handle dict-form algorithms in At Line 973, Suggested fix def _is_layerwise(obj):
if isinstance(obj, ModelOptPTQRecipe):
return _is_layerwise(obj.quantize.algorithm)
+ if isinstance(obj, dict):
+ if "layerwise" in obj:
+ return bool(obj["layerwise"])
+ if "algorithm" in obj:
+ return _is_layerwise(obj["algorithm"])
+ return False
if isinstance(obj, list):
return any(_is_layerwise(a) for a in obj)
return bool(getattr(obj, "layerwise", False))🤖 Prompt for AI Agents |
||
|
|
||
| is_layerwise = _is_layerwise(recipe) | ||
|
|
||
| if args.batch_size == 0: | ||
| # For VL models with image-text calibration, skip automatic batch size detection | ||
| # since get_max_batch_size can't handle multimodal inputs | ||
|
|
@@ -968,6 +987,11 @@ def quantize_main( | |
| "Offline speculative decoding calibration enabled. Using default batch_size=1 for calibration." | ||
| ) | ||
| args.batch_size = 1 | ||
| # Layerwise calibration processes one layer at a time; auto batch-size probing runs a | ||
| # full-model forward which defeats the point and can OOM on very large models. | ||
| elif is_layerwise: | ||
| print("Layerwise calibration enabled. Using default batch_size=1 for calibration.") | ||
| args.batch_size = 1 | ||
| else: | ||
| # Calibration/sparsification will actually take much more memory than regular inference | ||
| # due to intermediate tensors for fake quantization. Setting sample_memory_usage_ratio | ||
|
|
@@ -1027,12 +1051,7 @@ def quantize_main( | |
| else: | ||
| # mono quantization | ||
|
|
||
| if args.recipe is not None: | ||
| print(f"Use recipe {args.recipe} for quantization") | ||
| recipe = load_recipe(args.recipe) | ||
| assert isinstance(recipe, ModelOptPTQRecipe), ( | ||
| f"Expected PTQ recipe, but got {type(recipe).__name__} from {args.recipe}" | ||
| ) | ||
| if recipe is not None: | ||
| quant_cfg = recipe.quantize.model_dump() | ||
|
|
||
| else: | ||
|
|
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,94 @@ | ||
| # SPDX-FileCopyrightText: Copyright (c) 2026 NVIDIA CORPORATION & AFFILIATES. All rights reserved. | ||
|
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Okay for now. Can you compose this yaml based on the |
||
| # SPDX-License-Identifier: Apache-2.0 | ||
| # | ||
| # Licensed under the Apache License, Version 2.0 (the "License"); | ||
| # you may not use this file except in compliance with the License. | ||
| # You may obtain a copy of the License at | ||
| # | ||
| # http://www.apache.org/licenses/LICENSE-2.0 | ||
| # | ||
| # Unless required by applicable law or agreed to in writing, software | ||
| # distributed under the License is distributed on an "AS IS" BASIS, | ||
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
| # See the License for the specific language governing permissions and | ||
| # limitations under the License. | ||
|
|
||
| metadata: | ||
| recipe_type: ptq | ||
| description: NVFP4 static weight and dynamic activation for expert layers only (W4A4), FP8 KV cache, max layerwise calibration. | ||
| quantize: | ||
| algorithm: | ||
| method: max | ||
| # Max calibration is fast and does not typically need checkpointing. | ||
|
realAsma marked this conversation as resolved.
|
||
| layerwise: true | ||
| quant_cfg: | ||
| - quantizer_name: '*' | ||
| enable: false | ||
| - quantizer_name: '*mlp.experts*weight_quantizer' | ||
| enable: true | ||
| cfg: | ||
| block_sizes: | ||
| -1: 16 | ||
| type: dynamic | ||
| scale_bits: e4m3 | ||
| num_bits: e2m1 | ||
| - quantizer_name: '*mlp.experts*input_quantizer' | ||
| enable: true | ||
| cfg: | ||
| block_sizes: | ||
| -1: 16 | ||
| type: dynamic | ||
| scale_bits: e4m3 | ||
| num_bits: e2m1 | ||
| - quantizer_name: '*block_sparse_moe*weight_quantizer' | ||
| enable: true | ||
| cfg: | ||
| block_sizes: | ||
| -1: 16 | ||
| type: dynamic | ||
| scale_bits: e4m3 | ||
| num_bits: e2m1 | ||
| - quantizer_name: '*block_sparse_moe*input_quantizer' | ||
| enable: true | ||
| cfg: | ||
| block_sizes: | ||
| -1: 16 | ||
| type: dynamic | ||
| scale_bits: e4m3 | ||
| num_bits: e2m1 | ||
| - quantizer_name: '*[kv]_bmm_quantizer' | ||
| enable: true | ||
| cfg: | ||
| num_bits: e4m3 | ||
| - quantizer_name: '*block_sparse_moe.gate*' | ||
| enable: false | ||
| - quantizer_name: '*linear_attn.conv1d*' | ||
| enable: false | ||
| - quantizer_name: '*lm_head*' | ||
| enable: false | ||
| - quantizer_name: '*mixer.conv1d*' | ||
| enable: false | ||
| - quantizer_name: '*mlp.gate.*' | ||
| enable: false | ||
| - quantizer_name: '*mlp.shared_expert_gate.*' | ||
| enable: false | ||
| - quantizer_name: '*output_layer*' | ||
| enable: false | ||
| - quantizer_name: '*proj_out.*' | ||
| enable: false | ||
| - quantizer_name: '*router*' | ||
| enable: false | ||
| - quantizer_name: 'output.*' | ||
| enable: false | ||
| - parent_class: 'nn.BatchNorm1d' | ||
| quantizer_name: '*' | ||
| enable: false | ||
| - parent_class: 'nn.BatchNorm2d' | ||
| quantizer_name: '*' | ||
| enable: false | ||
| - parent_class: 'nn.BatchNorm3d' | ||
| quantizer_name: '*' | ||
| enable: false | ||
| - parent_class: 'nn.LeakyReLU' | ||
| quantizer_name: '*' | ||
| enable: false | ||
Uh oh!
There was an error while loading. Please reload this page.