Viewing a single comment thread. View all comments

cbsudux t1_jd1qzp7 wrote

How long did the training take on an A100?

1

benfavre t1_jd2n1cg wrote

1 epoch of finetuning the 30B model with llama-lora implementation, mini-batch-size=2, maxlen=384, is about 11 hours.

3