Submitted by New_Yak1645 t3_11nhl03 in MachineLearning

Most AI is impossible to train(like chat GPT)

Dose LLaMa can be trained?

Although the dataset is very hard to get, It would be nice if LLaMa can be trained.

When searching for reddit, this topic cannot be searched, so I hope it becomes a discuss about HW or availability.
Thank you.

0

Comments

You must log in or register to comment.

CKtalon t1_jbnccl7 wrote

If you have a few thousand A100s, sure? The dataset is fairly easily obtainable.

The next difficulty is the technical knowhow to train such LLMs.

16

ch9ki7 t1_jbneot3 wrote

I would start searching on huggingface.co

2

UnusualClimberBear t1_jbngux4 wrote

Training from scratch required 2048 A100 for 21 days. And it seems only to be the final run.

I guess you can start to fine-tune it with much lower resources, 16 A100 seems reasonable as going lower will require quantization or partial loadings for the model.

7

Raise_Fickle t1_jc1tb4r wrote

Any idea for finetuning llama on multi-gpu setup?

1