Disastrous_Elk_6375
Disastrous_Elk_6375 t1_jawatbb wrote
Reply to comment by WittyBananaPeel in Did you get access to Meta AI's LLAMA? [Discussion] by WittyBananaPeel
As a large language model I can't comment, but my friends say it works :D
Disastrous_Elk_6375 t1_javwm8d wrote
laughs in magnet link
Disastrous_Elk_6375 t1_jala1ee wrote
Wasn't this settled once and for all when they had the same exact model he worked on claim (very convincingly) that it was a frog and doing frog things, or something like this?
Disastrous_Elk_6375 t1_j9z7xvx wrote
Reply to comment by marketrent in Euclid space telescope launch scheduled for July — ESA mission to chart a 3D map of the universe, in search of dark matter and dark energy by marketrent
> "By subtracting the visible matter, we can calculate the presence of the dark matter which is in between," [Euclid project manager] Racca said.
This reminds me of this great nugget brought to us by the department of redundancy department:
The missile knows where it is at all times. It knows this because it knows where it isn't. By subtracting where it is from where it isn't, or where it isn't from where it is (whichever is greater), it obtains a difference, or deviation. The guidance subsystem uses deviations to generate corrective commands to drive the missile from a position where it is to a position where it isn't, and arriving at a position where it wasn't, it now is. Consequently, the position where it is, is now the position that it wasn't, and it follows that the position that it was, is now the position that it isn't. In the event that the position that it is in is not the position that it wasn't, the system has acquired a variation, the variation being the difference between where the missile is, and where it wasn't. If variation is considered to be a significant factor, it too may be corrected by the GEA. However, the missile must also know where it was.
Disastrous_Elk_6375 t1_j9yzjmx wrote
Reply to comment by HildemarTendler in Alien hunters get a boost as AI helps identify promising signals from space by UniOfManchester
No. There are definitely areas where ML can help. We have models that are known to be good at classification and that also generalise reasonably well. These models can and should be used to speed up the "anomaly detection" in a large amount of data. These models are also better at the task than manually defined "traditional" algorithms.
Disastrous_Elk_6375 t1_j9o0ctj wrote
Reply to comment by Bewaretheicespiders in Relativity Space on Twitter: You’ve asked, “Wen Launch?” and to that, we say...👇 Catch us live at Launch Complex 16 in Cape Canaveral, FL on March 8, 2023 to watch the world’s first 3D printed rocket fly. 🚀 #GLHF by allforspace
This is how SpX started ~20 years ago, and look at them now.
Disastrous_Elk_6375 t1_j9nrm6w wrote
Reply to comment by currentscurrents in [R] Provable Copyright Protection for Generative Models by vyasnikhil96
> It does memorize short snippets in some cases (especially when a snippet is repeated many times in training data)
And, to be fair, how can it not? How many different ways can you write a simple for loop to print some objects, or match a regex, call an API, and so on?
Disastrous_Elk_6375 t1_j9nr09q wrote
Reply to comment by kalenxy in Researchers discover mysterious source of 'heartbeat-like' radio bursts in a solar fare by AbbydonX
In CS a "heartbeat" is a periodical signal that is sent to confirm that a process is still running.
Disastrous_Elk_6375 t1_j9a2877 wrote
Reply to comment by ArmagedonAshhole in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
Thanks!
Disastrous_Elk_6375 t1_j99xxfa wrote
Reply to comment by ArmagedonAshhole in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
Are there some rough numbers on prompt size vs. ram usage after the model load? I haven't played yet with GPT-NeoX
Disastrous_Elk_6375 t1_j99ujv1 wrote
Reply to comment by head_robotics in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
add this to your .from_pretrained("model" , device_map="auto", load_in_8bit=True)
Transformers does the rest.
Disastrous_Elk_6375 t1_j99ry6s wrote
Reply to [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
GPT-NeoX should fit in 24GB VRAM with 8bit, for inference.
I managed to run GPT-J 6B on a 3060 w/ 12GB and it takes about 7.2GB of VRAM.
Disastrous_Elk_6375 t1_j8hdb2r wrote
Reply to comment by Zondartul in [R] [P] OpenAssistant is a fully open-source chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so. by radi-cho
Do you know if distilling will be possible after instruct finetuning and the RLHF steps? I know it works on "vanilla" models, but I haven't searched anything regarding distillation of instruct trained models.
Disastrous_Elk_6375 t1_j8cd4x4 wrote
Reply to comment by radi-cho in [R] [P] OpenAssistant is a fully open-source chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so. by radi-cho
I think it will depend on how small the LLMs that it uses are. If they can be run on consumer GPUs, then it will probably take off. If you need to rent 8xGPU servers just for inference, probably not.
Stablediffusion took off because in the first two weeks you could run it on 4GB VRAM GPUs. Then when "finetuning" aka dreambooth came along, it went from 24 to 16 to 8 GB in a matter of weeks. Same effect there.
Disastrous_Elk_6375 t1_j885wd6 wrote
Check out this - https://huggingface.co/models
You can download models and try them out locally, depending on your specs. It's unlikely you'll find a single model that does everything you need, but there's a chance you can use a combination of models to get close to what you want. You'll need to be a bit more specific about your end goals to get better suited suggestions.
Disastrous_Elk_6375 t1_j87rwu5 wrote
Reply to comment by Maximum-Geologist-98 in [D] Have their been any attempts to create a programming language specifically for machine learning? by throwaway957280
As a large language model I have to caution against using sharp objects in programming languages, as it would pose a great risk to the developers unknowingly hurting themselves with it. Furthermore, it can be said that axes are typically not very sharp, and as we know blunt objects are objects that are not very sharp and also might not be extremely sharp. Sharp is a now defunct company that used to produce TV sets. A TV set is like a modern TV but it used to also be old. /s?
Disastrous_Elk_6375 t1_j836fxc wrote
Reply to comment by glhope in SpaceX on Twitter: Super Heavy Booster 7 completed a full duration static fire test of 31 Raptor engines, producing 7.9 million lbf of thrust (~3,600 metric tons) – less than half of the booster’s capability by allforspace
They're testing engines separately at the factory. They've ran hours of tests and most likely have a pretty solid understanding of what thrust each engine gives at a certain "throttle" level. So they'll have precise measurements of things like flow for each engine, and they'll know what each flow setting would translate into thrust. From there it's simple math and some approximation.
Disastrous_Elk_6375 t1_j809rd6 wrote
Reply to comment by AreEUHappyNow in Blue Origin awarded NASA launch contract for Mars mission (Studying magnetic field) by kuroimakina
> SpaceX in particularly has also paid Russia for access to Soviet rocketry,
They what now?
Disastrous_Elk_6375 t1_j7hvc22 wrote
Reply to comment by Mother-Wasabi-3088 in Rolls-Royce Nuclear Engine Could Power Quick Trips to the Moon and Mars by darthatheos
> Nuclear power will never be safe
Mmmhhmm. We've had extremely safe, sufficiently compact and mobile nuclear power since the 50s. We know they're safe because navy personnel on nuclear subs / ships have lived long healthy lives. In fact, the commander of the first US nuclear sub (commissioned in 1954) went on to also command the first nuclear ship. He got to live 94 years!
Disastrous_Elk_6375 t1_j2m49ih wrote
Reply to comment by avatarOfIndifference in [D] Data cleaning techniques for PDF documents with semantically meaningful parts by cm_34978
Oh yeah. "There are several", "some of the most ... include", "it may be necessary", "overall the best approach" are 100% markers of chatgpt that I've seen in most answers that I got.
Disastrous_Elk_6375 t1_j2ft5zo wrote
Reply to comment by Glycerine in An Open-Source Version of ChatGPT is Coming [News] by lambolifeofficial
> You're right it's poor. All 8 CPU's hit 100%.
Yeah, you're probably not using the gpu. Make sure that your pytorch & cuda stuff are compatible and properly installed. To test, go into a python session, and do
torch.cuda.is_available()
If the output is false it will train on CPU.
Disastrous_Elk_6375 t1_j2e6d6d wrote
Reply to comment by Glycerine in An Open-Source Version of ChatGPT is Coming [News] by lambolifeofficial
> 92.98s/it
Are your CPUs fully used when training? You might want to check if this is running on GPU or not, those numbers are generally found on CPU training.
Disastrous_Elk_6375 t1_j2de4o2 wrote
Reply to comment by ThatInternetGuy in An Open-Source Version of ChatGPT is Coming [News] by lambolifeofficial
Can the 4090 pool their VRAM? I always thought that LLMs need GPUs from the A/V series so that they can pool memory. Am I wrong in thinking that?
Disastrous_Elk_6375 t1_jb8y5r2 wrote
Reply to [D] Neat project that would "fit" onto a 4090? by lifesthateasy
GptNeoX should fit with 8bit and low prompt sizes. GptJ-7B should fit as well with 16bit inference. On smaller models you might even be able to do some finetuning for fun.
There's a couple of coding models from salesforce that you could fit comfortably. Check out FauxPilot for a copilot "clone".