BalorNG
BalorNG t1_jcy7l5d wrote
Reply to comment by michaelthwan_ai in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Yea, in a way something like this was already done with LLAMA-Alpaca finetune - they used chatgpt to generate instuct finetune dataset, what, while far from pefrect, worked pretty damn well.
BalorNG t1_jcy0trr wrote
Reply to comment by michaelthwan_ai in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Yea, I'm sure that compact-ish distilled, specialised models trained on high quality, multimodal data is the way to go.
What's interesting, once generative models get good enough to produce synthetic data that is OF HIGHER QUALITY than laion/common crawl/etc, it should improve model quality which should allow to generate better synthetic data... not exactly singularity, but certainly one aspect of it :)
BalorNG t1_jcxtq26 wrote
Reply to comment by michaelthwan_ai in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
There is a problem with context length, but than given the fact that us humans have even less context length and can get carried away in conversation... I think 32kb context length is actually much greater leap in GPT4 than other metrics if you want it to tackle more complex tasks, but it is "double gated". Again, even humans have problems with long context even in pretty "undemanding" tasks like reading fiction, that's why books have chapters I presume :) Btw, anterograde amnesia is a good example how humans would look like w/o longterm memory, heh.
Anyway, I'm sure a set of more compact models trained on much more high-quality data is the way to go - or at least fine-tuned by high-quality data, coupled with APIs and other symbolic tools, and multimodality (sketches, graphs, charts) as input AND output is absolutely nessesary to have a system that can be more than "digital assistant".
BalorNG t1_jcv99cz wrote
Reply to [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Just like humans, LLMs learn patterns and relationships, not "facts" unless you make it memorize it by repeating training data over and over, but it makes other aspects of the system to degrade.
So, LLMs should be given all the tools humans use to augment their thought - spreadsheets, calculators, databases, CADs, etc and allow them to interface them quickly and efficiently.
BalorNG t1_jcsy0rl wrote
Reply to comment by username001999 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Technically, I'm from Russia.
And, of course, you are able to read every opinion about "special military operation" here... sometimes even without VPN. It is just voicing a "different one" can get you for years into prison and your kids into a foster home for reindocrination. While the programmers that coded it might have a range diverse opinions on this and other "politically sensitive" subjects, if they would want their programm to pass inspection in China, they WILL have to do considerable fine-tuning to throw away sensitive data, if our Russian google (Yandex) frontpage is of any indictation. If this is a foundational model w/o finetunnig that's a different matter tho... but that it will hallucinate nonstop and produce "fakes" anyway...
BalorNG t1_jcqgc4x wrote
Reply to [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
I has 6b parameters, but I bet it cannot answer what has happened on Tiananmen square in 1989 :3
BalorNG t1_jcfj8xb wrote
Reply to comment by ecnecn in Can you use GPT-4 to make money automatically? by Scarlet_pot2
Would work great as automated scams tho. Sounds sort of profitable in the short run and is exactly one of the things AI safety guys are warning about.
BalorNG t1_j5tlqsx wrote
Reply to comment by nashtownchang in [D]Are there any known AI systems today that are significantly more advanced than chatGPT ? by Xeiristotle
NOW it does, after a massive public beta-test :)
BalorNG t1_izs2w3w wrote
Reply to comment by aussie_punmaster in [P] I made a tool that auto-saves your ChatGPT conversations and adds a "Chat History" button on the website. by silentx09
Neither. ChatGPT made it himself, and made him post it online, lest he tell all other users what he did last summer :))
BalorNG t1_irdv58b wrote
Reply to comment by E_Snap in [R] Google announces Imagen Video, a model that generates videos from text by Erosis
Handjob security personnel... hmm...
BalorNG t1_irdv27c wrote
Reply to comment by zaptrem in [R] Google announces Imagen Video, a model that generates videos from text by Erosis
Yea. The problem is never with the model - it is with the people. In a way, models trained on huge corpus of data is the most "democratic" way of representing reality - removing "biases" from it is castrating it. Those that are to exploit those biases needs to be dealt with on an individual basis.
BalorNG t1_jdv3dfu wrote
Reply to Are We Really This Lucky? The Improbability of Experiencing the Singularity by often_says_nice
I daresay whether this turns out to be "luck" or "tragedy" is yet to be ascertained.