fix(docs): fix typos in comments and docstrings (#34836)

Signed-off-by: machov <mv1742@nyu.edu>
This commit is contained in:
Manrique Vargas
2026-02-19 02:17:41 -05:00
committed by GitHub
parent 9681068cf9
commit ad5aa6bd9f
5 changed files with 6 additions and 6 deletions

View File

@@ -285,7 +285,7 @@ class CompilerManager:
with self.compile_context(compile_range):
# There is a compilation time optimization here.
#
# If the (input metdata, graph, compiler config) are the same, then
# If the (input metadata, graph, compiler config) are the same, then
# we want to avoid compiling the same artifact again. If we didn't
# do this optimization, the backend compilation (InductorAdaptor or
# InductorStandaloneAdaptor)

View File

@@ -420,7 +420,7 @@ def make_fp8_moe_quant_config(
per_out_ch_quant: bool = False,
) -> FusedMoEQuantConfig | None:
"""
Create FusedMoEQuantConfig for the specifed FP8 Backend.
Create FusedMoEQuantConfig for the specified FP8 Backend.
The FusedMoEQuantConfig holds the scales that are used
at runtime by the Modular Kernel abstraction.

View File

@@ -151,7 +151,7 @@ class DefaultMoERunner(MoERunner):
kernels for different parallel execution modes.
Eventually, this class will be split up and specialized for different
configurations, e.g. the presense or absence of shared experts, a gate, etc.
configurations, e.g. the presence or absence of shared experts, a gate, etc.
"""
def __init__(

View File

@@ -586,7 +586,7 @@ class GptOssModel(nn.Module):
parts = name.split(".")
ids = [s for s in parts if s.isdigit()]
# for amd-quark format that each expert is seperated
# for amd-quark format that each expert is separated
# need to extract the parameter name with experts fused.
# example model: amd/gpt-oss-20b-MoE-Quant-W-MXFP4-A-FP8-KV-FP8
if len(ids) == 2:

View File

@@ -567,8 +567,8 @@ def current_stream() -> torch.cuda.Stream:
return _current_stream_tls.value
# Global auxilary stream for running operations in background streams.
# We have single global auxilary stream to avoid an explosion of streams
# Global auxiliary stream for running operations in background streams.
# We have single global auxiliary stream to avoid an explosion of streams
# for every layer (and make profiling look sane).
#
# aux_stream() is currently used for: