Skip to content

Commit

Permalink
Update fp8.py
Browse files Browse the repository at this point in the history
  • Loading branch information
Yantom1 authored Sep 25, 2024
1 parent f58d4c1 commit db9affe
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion vllm/model_executor/layers/quantization/fp8.py
Original file line number Diff line number Diff line change
Expand Up @@ -120,7 +120,7 @@ class Fp8LinearMethod(LinearMethodBase):

def __init__(self, quant_config: Fp8Config):
self.quant_config = quant_config
if current_platform.is_cuda_alike()
if current_platform.is_cuda_alike():
self.cutlass_fp8_supported = cutlass_fp8_supported()

# For GPUs that lack FP8 hardware support, we can leverage the
Expand Down

0 comments on commit db9affe

Please sign in to comment.