During quantization in the Fargan module, at line 184 of common.py, scale_sum = weight_max_sum / 129- I'm not sure why this conversion needs to be configured with 129.Additionally, I'd like to know if there exists any code related to int16 and f16 quantization?
During quantization in the Fargan module, at line 184 of common.py, scale_sum = weight_max_sum / 129- I'm not sure why this conversion needs to be configured with 129.Additionally, I'd like to know if there exists any code related to int16 and f16 quantization?