Submitted by Angry_Grandpa_ t3_y92cl1 in singularity
Angry_Grandpa_ OP t1_it4qmjp wrote
Reply to comment by manOnPavementWaving in A YouTube large language model for a scant $35 million. by Angry_Grandpa_
It's based on the Chinchilla paper and not my personal opinion. You should read the paper.
If you think the assumptions are wrong you should do your own projections based on the paper.
manOnPavementWaving t1_it4r8la wrote
I have read the paper, which is how I know that they scale data and parameters equally, meaning a 10x in data results in a 100x in compute required and hence a 100x in cost.
Assumptions wise Im looking more at the number of words on youtube, your estimate is likely wildly off.
Youre also ignoring that the training time could very well be long enough that it would be a better strategy to wait for better GPUs to come out.
Angry_Grandpa_ OP t1_it50h5j wrote
What is your estimate?
LeroyJanky80 t1_it6g9ub wrote
The data Google/Alphabet has is obviously it's most powerful asset. My guess is they've done this, have the means, brain trust, wealth and capacity to do this. They can easily cover this in all domains where people, infrastructure and content are concerned. It's a massive endeavour but so is what they did with the entire internet many many years ago and at the time it was groundbreaking.
Viewing a single comment thread. View all comments