benfavre
benfavre t1_jd2n1cg wrote
Reply to comment by cbsudux in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
1 epoch of finetuning the 30B model with llama-lora implementation, mini-batch-size=2, maxlen=384, is about 11 hours.
benfavre t1_jedx7pb wrote
Reply to [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
It's a pity that neither weights nor training data are made available.