torch_dtype only for torch.float16?
yumianhuli1 opened this issue · comments
yumianhuli commented
Does inference currently only support torch_dtype=torch.float16?int8_float16、int8 will be supported?
yumianhuli1 opened this issue · comments
Does inference currently only support torch_dtype=torch.float16?int8_float16、int8 will be supported?