manOnPavementWaving
manOnPavementWaving t1_it393xz wrote
my man you cant just scale cost with number of tokens and not number of parameters
way too many mostly false assumptions in these calculations
manOnPavementWaving t1_issmqzx wrote
Reply to comment by Redifyle in Talked to people minimizing/negating potential AI impact in their field? eg: artists, coders... by kmtrp
I don't think this holds for higher layers of abstraction. If all I need to do is ask to make the porgram faster, and then a model does it for me, that's easy. If I ask it to optimize cache hits and data locality, that task would be more difficult for the prompter (more specific). Depending on the quality of the system, the level of abstraction will eventually reach a point where anyone can code, with very little background knowledge.
manOnPavementWaving t1_isqdf9x wrote
Reply to comment by Owner2229 in Talked to people minimizing/negating potential AI impact in their field? eg: artists, coders... by kmtrp
Disagree, gave my dad my DALL-E access, and he learned to prompt engineer to reasonable success in an hour. Once you've got quick iteration, specifying what you want becomes trivial.
manOnPavementWaving t1_isqd6kk wrote
Reply to comment by NoRip7374 in Talked to people minimizing/negating potential AI impact in their field? eg: artists, coders... by kmtrp
Not my point, the point was if you've genuinely automated software creation, you've almost immediately automated everything else
manOnPavementWaving t1_isp7v7u wrote
Reply to Talked to people minimizing/negating potential AI impact in their field? eg: artists, coders... by kmtrp
As a coder, I feel quite safe. Not because Im denying progress, but because if Im not safe from automation, nobody is. Making me quite safe.
manOnPavementWaving t1_ir9eo4d wrote
Reply to comment by superluminary in The last few weeks have been truly jaw dropping. by Particular_Leader_16
Thats my entire point😅
manOnPavementWaving t1_ir7synf wrote
Reply to comment by Dr_Singularity in Turns out everything is a matrix multiplication from computer graphics to training neural networks :) Our latest front cover Nature paper on AlphaTensor by Dr_Singularity
I feel like pretty soon this is gonna feel like "lucky nature", and not "lucky deepmind", if it doesn't already feel like that
Thank god that, at least AI people, dont care too much about journals, and more about functional and/or elegant ideas
manOnPavementWaving t1_ir7sg8u wrote
I still cant grasp the reason behind "AI is progressing fast, so singularity will happen this decade". Maybe it will, but without a list of things/milestones needed for the singularity and reasonable estimates for each of them for when we're gonna reach them, (none of which you can completely defend because we don't actually know), such estimates hardly have any degree of certainty.
manOnPavementWaving t1_iqv3mmi wrote
Reply to comment by 2Punx2Furious in Large Language Models Can Self-improve by Dr_Singularity
Its in the author's best interests to show of who they are, misaligning that tends to just result in subtly cheating the system
Peer review in AI has been less and less important though, trial by twitter tends to perform much better
manOnPavementWaving t1_iqutco3 wrote
Reply to comment by Sashinii in Is ai countdown accurate? by Phoenix5869
Realistic being a relative term here, Im more on Sam Altman's side who said not too long ago it'll be about 7 years before AI can give a good TED talk completely on its own
manOnPavementWaving t1_iqtkra0 wrote
Reply to comment by Nmanga90 in Large Language Models Can Self-improve by Dr_Singularity
Actually we know what the LM is, it's PaLM, developed by google under Jeff Dean.
Anonymous peer review is a fucking joke
manOnPavementWaving t1_iqq9xxp wrote
Reply to comment by adt in Dramatron: Co-Writing Screenplays and Theatre Scripts with Language Models (DeepMind) by nick7566
I meannn, they had probably written most of the paper by then
manOnPavementWaving t1_iqohe5w wrote
Reply to comment by Midori_Schaaf in Why I am optimistic about the Optimus bot by Effective-Dig8734
Except of course for all the companies that have also been doing this. Saycan (https://arxiv.org/abs/2204.01691) actually manages to perform arbitrary tasks with very decent performance, more transparent research (not in open sourcing, but definitely in explaining their method) and google has much more experience with AI (not that it matters much).
Optimus could become very cool, but Tesla decidedly hasnt got an edge in this space
manOnPavementWaving t1_it4r8la wrote
Reply to comment by Angry_Grandpa_ in A YouTube large language model for a scant $35 million. by Angry_Grandpa_
I have read the paper, which is how I know that they scale data and parameters equally, meaning a 10x in data results in a 100x in compute required and hence a 100x in cost.
Assumptions wise Im looking more at the number of words on youtube, your estimate is likely wildly off.
Youre also ignoring that the training time could very well be long enough that it would be a better strategy to wait for better GPUs to come out.