Skip to content
This repository has been archived by the owner on Dec 1, 2024. It is now read-only.

question about quantization #119

Open
xinhaoc opened this issue Jun 4, 2023 · 0 comments
Open

question about quantization #119

xinhaoc opened this issue Jun 4, 2023 · 0 comments

Comments

@xinhaoc
Copy link

xinhaoc commented Jun 4, 2023

Hi FlexGen team! I have a question about your quantization algorithm. are you using this function run_float_quantization for int4/int8 compression? When I run the test(test_float_quantize), it fails because the params is different with the deepspeed version(the ref_out_tensor is the same). the deepspeed param can recover the float16 tensor, the run_float_quantize can't. Thanks!

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant