Hi Everyone, I am trying to quantize my model to 4-bit using post-training quantization, I see tflite supports int 8 bits and float 16 bit. How can do post-training quantization to convert my float 32-bit model to int 4-bit model . Is it possible ? Kindly provide your views
Regards
Saras