Closed
Description
Hi,
-
When playing with MPT-7b models, I frequently meet the issues of "attn_bias not aligned", with tensor_parallel_size - 2, how do alleviate this issue?
-
Besides, I just find that your default model loading scripts load float16 versions, for fair evaluation, is it necessary to switch to float32?
Thanks very much in advance!