incorrect use of T5 model?
#39
by
X-niper
- opened
We compare the fp16 T5 model in this repo and the fp32 T5 model with the same parameters and find that the outputs for the same text prompt are not the same.
Possible solution may be: cast the model to fp32 and do inference with autocast(bf16) or directly with fp32.