[Kernel] Re-tune Mixtral MoE configurations for FP8 on H100 (#5238)
This commit is contained in:
@@ -255,7 +255,8 @@ def main(args: argparse.Namespace):
|
||||
|
||||
if args.batch_size is None:
|
||||
batch_sizes = [
|
||||
1, 2, 4, 8, 16, 32, 64, 128, 256, 512, 1024, 1536, 2048, 3072, 4096
|
||||
1, 2, 4, 8, 16, 24, 32, 48, 64, 96, 128, 256, 512, 1024, 1536,
|
||||
2048, 3072, 4096
|
||||
]
|
||||
else:
|
||||
batch_sizes = [args.batch_size]
|
||||
|
||||
Reference in New Issue
Block a user