Skip to content

Commit d339190

Browse files
Layali Rashidmmarcinkiewicz
authored andcommitted
Add fp8 attn knobs
1 parent bf9182a commit d339190

File tree

2 files changed

+4
-1
lines changed

2 files changed

+4
-1
lines changed

megatron/core/fp4_utils.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,9 @@ def get_fp4_recipe(config: TransformerConfig):
6363
if is_te_min_version("2.7.0.dev0"):
6464
if config.fp4_recipe == Fp4Recipe.nvfp4:
6565
try:
66-
fp4_recipe = transformer_engine.common.recipe.NVFP4BlockScaling()
66+
fp4_recipe = transformer_engine.common.recipe.NVFP4BlockScaling(
67+
fp8_dpa=config.fp8_dot_product_attention,
68+
)
6769
except AttributeError:
6870
raise ValueError(
6971
"""NVFP4BlockScaling recipe is not available in this version of

megatron/core/fp8_utils.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -437,6 +437,7 @@ def get_fp8_recipe(config: TransformerConfig):
437437
)
438438
elif config.fp8_recipe == Fp8Recipe.tensorwise and is_te_min_version("2.2.0.dev0"):
439439
fp8_recipe = transformer_engine.common.recipe.Float8CurrentScaling(
440+
fp8_dpa=config.fp8_dot_product_attention,
440441
fp8_format=fp8_format
441442
)
442443
elif config.fp8_recipe == Fp8Recipe.blockwise and is_te_min_version("2.3.0.dev0"):

0 commit comments

Comments
 (0)