CellWithoutCulture
CellWithoutCulture t1_jcmsxjq wrote
Reply to comment by ThatInternetGuy in Those who know... by Destiny_Knight
HF-RLHF is the name of the dataset. As far as RLHF... what they did to LLaMA is called "Knowledge Distillation" and iirc usually isn't quite as good as RLHF. It's an approximation.
CellWithoutCulture t1_jcjkycz wrote
Reply to comment by BSartish in Those who know... by Destiny_Knight
decent video
CellWithoutCulture t1_jcjkwy1 wrote
Reply to comment by ThatInternetGuy in Those who know... by Destiny_Knight
Most likely they haven't had time.
They can also use SHP and HF-RLHF.... I think they will help a lot since LLaMA didn't get the privlidge of reading reddit (unliked ChatGPT)
CellWithoutCulture t1_jcjku3z wrote
Reply to comment by Yomiel94 in Those who know... by Destiny_Knight
The specific type of fine-tuning was called Knowledge Distillation, I believe. ChatGPT taught LLaMA to chat, "stealing" OpenAI's business edge in the process.
CellWithoutCulture t1_javqw9s wrote
Reply to comment by LetterRip in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Fantastic reply, it's great to see all those concrete advances thst made it intro prod. Thanks for sharing.
CellWithoutCulture t1_javhjpc wrote
Reply to comment by LetterRip in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
I mean... why were they not doing this already? They would have to code it but it seems like low hanging fruit
> memory efficient attention. 10x-20x increase in batch size.
That seems large, which paper has that?
CellWithoutCulture t1_ja7dklj wrote
Reply to comment by tdgros in [D] Is RL dead/worth researching these days? by [deleted]
> Toolformer
....oh you're right it didn't. I assumed they let it use any tool which would need RL. But it seems like they had pre-labelled ways to use tools.
Thanks for pointing that out.
CellWithoutCulture t1_ja6pjet wrote
Seems more like an AskML question.
But RL is for situations when you can't backprop the loss. It's noisier than supervised learning. So if you can use supervised learning, then that's what you should generally use.
RL is still used, for example the recent GATO and Dreamer v3. Or used in training an LLM to use tools like in toolformer. And also OpenAI's famous RLHF, which stands for reinforcement learning with human feedback. This is what they use to make ChatGPT "aligned" although in reality it doesn't get there.
CellWithoutCulture t1_ja5tfhe wrote
There's a novel about his: Walkaway but Cory Docterow
CellWithoutCulture t1_j9syug9 wrote
Reply to comment by [deleted] in What are the big flaws with LLMs right now? by fangfried
ok stop it now
CellWithoutCulture t1_j9rbqdg wrote
Reply to comment by Present_Finance8707 in What are your thoughts on Eliezer Yudkowsky? by DonOfTheDarkNight
> kiddo
Wow you must be fun at parties
CellWithoutCulture t1_j9nqm32 wrote
Reply to comment by Molnan in What are your thoughts on Eliezer Yudkowsky? by DonOfTheDarkNight
This guy is being weird dismissing you for not knowing gwern.
But I will add that Gwern is awesome and a pleasure to read. You will probably find insight in reading and skimming some of his stuff.
e.g.
CellWithoutCulture t1_j9nploc wrote
Reply to comment by DonOfTheDarkNight in What are your thoughts on Eliezer Yudkowsky? by DonOfTheDarkNight
That's not right. He is a researcher.
You know Bostrom right? He wrote the book on the subject and credits Yud for many things. I mean what more do you want? Influential essays? Theories? They are all out there.
He's just no much of a coder from what I can see. Having released little code.
CellWithoutCulture t1_j9norym wrote
Reply to comment by Present_Finance8707 in What are your thoughts on Eliezer Yudkowsky? by DonOfTheDarkNight
And if you don't know who Drexler is...
I know who all these people are, yet I don't know anything lol
CellWithoutCulture t1_j9noid8 wrote
Reply to comment by Kolinnor in What are your thoughts on Eliezer Yudkowsky? by DonOfTheDarkNight
Yeah the jargon and meta rambling is so annoying. It's like their first priority is to show off their brains, and their second priority is to align AGI. Now they are almost finished showing of their brains, so watch out AGI.
Sometime they behave in a silly fashion. Greek philosopher's had excellent logic and deduced all kinds of wrong things. These guys seem similar at times, trying to deduce everything with philosophy and mega brains. .
IMO they are at their best when it's said in short form and it's grounded by empirical data.
There is also a lesswrong podcast or two that will read out some of the longer stuff.
CellWithoutCulture t1_j8fiz1j wrote
Reply to comment by p3opl3 in Anthropic's Jack Clark on AI progress by Impressive-Injury-91
Maybe they are doing a raise soon. They hired Karparthy because he's good but also because his reputation will help with raising ,especially with the narrative of a critical mass of talent. I may even be true.
CellWithoutCulture t1_j8fi9eq wrote
Reply to comment by visarga in Anthropic's Jack Clark on AI progress by Impressive-Injury-91
Yeah I love the part where he described why it matters. It shows he really understood the paper and is filtering through the noise.
CellWithoutCulture t1_j8fgpii wrote
Reply to comment by Slapbox in Anthropic's Jack Clark on AI progress by Impressive-Injury-91
then inf, then nan
CellWithoutCulture t1_j8fgbpg wrote
Reply to comment by space_troubadour in Anthropic's Jack Clark on AI progress by Impressive-Injury-91
I think he mean it's now super-exponential. It's rising faster than an exponential curve.
CellWithoutCulture t1_j8fg6st wrote
Reply to comment by helliun in Anthropic's Jack Clark on AI progress by Impressive-Injury-91
I think he mean it's now super-exponential
CellWithoutCulture t1_j7ae37n wrote
There's a book called "rapture of the nerds" that explain what happens to the Amish. Read it, it's a work of rare rick and morty level genius.
CellWithoutCulture t1_j6gnrgy wrote
Reply to comment by YobaiYamete in ChatGPT creator Sam Altman visits Washington to meet lawmakers | In the meetings, Altman told policymakers that OpenAI is on the path to creating “artificial general intelligence,” by Buck-Nasty
Is the potato chip thing something really thinks? That subreddit is awesome, thanks for sharing
CellWithoutCulture t1_j5xfc06 wrote
Reply to comment by SoylentRox in This subreddit has seen the largest increase of users in the last 2 months, gaining nearly 30k people since the end of November by _dekappatated
What the activation function paper?
Neural architecture search is certainly helping, you see it with optimized models coming out. And I presume chip design is.
Crazy days!
CellWithoutCulture t1_j5rmex9 wrote
Reply to comment by Left-Shopping-9839 in This subreddit has seen the largest increase of users in the last 2 months, gaining nearly 30k people since the end of November by _dekappatated
Are they really saying consciousness will emerge from LLM's, rather than that intelligence may arise from them?
CellWithoutCulture t1_jcr9g0g wrote
Reply to An Appeal to AI Superintelligence: Reasons to Preserve Humanity by maxtility
If you want this to be included in the training corpus of future language models, please upvote it.
Why? Well, language models are trained on the pile and common crawl. How do these dataset decide what to include? They look at reddit upvotes for one.
So you can influence what language models see in their formative years. (although they might not look at this subreddit).