philipgutjahr
philipgutjahr t1_jdn95o8 wrote
Reply to comment by michaelthwan_ai in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
for completeness, you should also add all those proprietary models: Megatron-Turing (530B, NVIDIA), Gopher (280B, Google), Chinchilla (70B, DeepMind) and Chatgenie (WriteCream)
philipgutjahr t1_jdn2p2u wrote
Reply to comment by michaelthwan_ai in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
from https://www.reddit.com/r/MachineLearning/comments/11uk8ti/d_totally_open_alternatives_to_chatgpt/
OpenChatKit (based on GPT-NeoX-20B) https://www.together.xyz/blog/openchatkit
Instruct-GPT https://carper.ai/instruct-gpt-announcement/
philipgutjahr t1_jctbs35 wrote
Reply to comment by Smallpaul in [P] The next generation of Stanford Alpaca by [deleted]
which can make a huge difference: GPT-3 + RLHF = Chat-GPT
philipgutjahr t1_jct7e53 wrote
Reply to comment by [deleted] in [P] The next generation of Stanford Alpaca by [deleted]
LGTM
philipgutjahr t1_jdq43bq wrote
Reply to comment by michaelthwan_ai in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
not sure if this is true, but afaik chat-gpt is basically a implementation of instruct-gpt (where OpenAI have been very thoroughly at RLHF)
"instance of" https://nextword.dev/blog/chatgpt-instructgpt-gpt3-explained-in-plain-english
"sibbling but a lot better" https://openai.com/blog/chatgpt