BellyDancerUrgot

BellyDancerUrgot t1_j6zyiqm wrote

I’ll be honest, I don’t really know what FPGAs (I reckon they are an ASIC for matrix operations?) do and how they do it but tensor cores already provide optimization for matrix / tensor operations and fp16 and mixed precision has been available for quite a few years now. Ada and hopper even enable insane performance improvements for fp8 operations. Is there any real verifiable benchmark that compares training and inference time of the two?

On top of that there’s the obvious Cuda monopoly that nvidia has a tight leash on. Without software even the best hardware is useless and almost everything is optimized to run on Cuda backend.

0

BellyDancerUrgot t1_j6v9w0p wrote

Reply to comment by FastestLearner in Using Jupyter via GPU by AbCi16

Last I checked for tensorflow-gpu conda install didn’t install the correct cuda version for some reason and it was annoying to roll back and then reinstall correct cuda and cudnn versions. PyTorch is fking clean tho.

0

BellyDancerUrgot t1_j6bzv82 wrote

Reply to comment by gantork in Google not releasing MusicLM by Sieventer

Using scraped data for research does not violate copyright laws. Monetizing it as a product for the public does. Most of the work done by Meta , Google , nvidia and other big tech arent even available for public use let alone monetized for public use. But yeah sure whatever u say! I’ve realized people on this sub who have no real knowhow about ML/DL and about laws/legal consequences are the ones that are the loudest.

Have a good day.

1

BellyDancerUrgot t1_j67sn0r wrote

Reply to comment by CypherLH in Google not releasing MusicLM by Sieventer

It isn’t at all. ‘Lol’

What I understand from our brief exchange :

-u have no idea about fair use, Creative Commons licensing, TDM rules apply to non commercial uses which is not the case here, scraping copyright protected content is a legal infringement if used for commercial purposes and or generate profit.

-u make dumb analogies because u don’t understand that representations in DL are equivalent to a photocopy of ur data. U can’t remove an artists watermark and use their IP to generate revenue.

oh but I can look at someone’s work and modify it a bit and thats fair use - yes except that’s not what’s happening here. Stop trying to throw random analogies trying to connect the two. Ur ai generated art will have the same distribution as whatever input data it sourced from during inference. Which is the entire foundation for digital watermarking against generative diffusion and GAN models which picked up in popularity.

−3

BellyDancerUrgot t1_j67qd3w wrote

Reply to comment by CypherLH in Google not releasing MusicLM by Sieventer

Totally wrong. A neural network learns a representation from the data. It literally scans ur work. The entire analogy of it ‘just looking’ at ur data is wrong. There’s a reason why artists have watermarks and signatures on artwork hosted on various websites. Circumventing measures put in place to prevent misuse doesn’t mean it’s legal , it just means existing laws were inadequate.

Edit: fyi there’s already work being done to trace back datasets on which ai art generation models were trained. Quite easy to do since most GAN and Diffusion models have distributions that get replicated in the output (cuz the outputs are derived from the representations learnt from the dataset they are trained on) making them easy to trace back.

−1

BellyDancerUrgot t1_j67p0e9 wrote

A degree , maybe a part time professional masters perhaps from a school where the faculty who teaches does active research. Or just read papers. 2 min papers is a good channel to start off with.

In ML/DL 1 year is already ancient. 4 years is prehistoric lol. For context if u choose a topic, say 2D-3D translation , from the advent of NERFs a couple of years back? We have a stupid amount of papers on the topic trying various novel approaches , everything ranging from using Voxels to store geometry in new ways, to geometry aware GANs , multi view compression using ViTs etc etc

So choose a topic and focus on that otherwise it’s a lot.

4

BellyDancerUrgot t1_j67oidu wrote

Reply to comment by CypherLH in Google not releasing MusicLM by Sieventer

Yes but these models were trained on data publicly available without consent. That’s the big legal problem and imo entirely fair. Fair use falls flat in this argument lol.

Edit : for people replying to my last comment

First mistake is comparing neural networks to the brain under this context.

And no

Their output is not unique because it follows the same distribution that it learnt the representation on. Humans don’t do that. You can’t find a human analogy because humans do not learn things the same way as neural nets.

Neural networks can’t actually extrapolate data because they don’t have a physical intuition just a large associative memory. You only think they can because you are uneducated on the topic.

−2

BellyDancerUrgot t1_j3nq5pn wrote

There would be a 5-8% overhead for the same gpu in a bare vm vs physical comparison. A100 is significantly faster for ML workloads than a 3090 iirc. So it’s probably something related to how it’s setup in your case. Also try using a single gpu instead of distributed learning if you are. MPI might be leading to more overhead in your compute node.

2

BellyDancerUrgot t1_j3efn9s wrote

I don’t have a PhD either lol. Your beliefs aren’t meaningless either. Nobody actually knows what breakthrough we might have next. I do consider chatgpt to be a breakthrough tbh (using RL to train an LLM). VQA was a breakthrough imo. GANs was also a breakthrough. All these came about in the same way as the post suggests but without hardware or funding u would never see all of it come together.

There’re people like Blake Richards working on the boundaries of neuroscience and AI but it’s hard to work on any of those fields without math as the underlying structure. Still, even if you approach it or want to approach it from an entirely new way it’s hard to do that without knowing the approaches that do exist which would require you to have a lot of math knowledge regardless. You can do that without a degree for sure tho , that wasn’t my point. It’s just super hard without guidance and the primary topic of this post is: working on smaller problems without any funding , I don’t see how that works and i don’t see any actual pragmatic answers here by op either.

1

BellyDancerUrgot t1_j3dtr99 wrote

Do share then what your beliefs are. What exactly is AI without math? Just curious since you have the tag of a researcher. What field are u working on? I’m not suggesting that a PhD is necessary, a degree is a indicator of ur work. But phd level work is necessary to achieve anything meaningful in this field. The post is very hand wavy and aimless. Andrew Ng and Khan academy is not enough to invent the next big thing however small it is. Read up on Mish activation. The guy who did that did so before even getting a masters degree. But that’s only because he is a genius who was capable of understanding grad math when barely out of high school.

1

BellyDancerUrgot t1_j311o8o wrote

No because humans do not hallucinate information and can derive conclusions based on cause and effect on subjects it hasn’t seen before. LLMs can’t even differentiate between cause and effect without memorizing patterns, something humans can naturally do.

And no, human beings in fact do not parrot information. I can reason about subjects I have never studied because human beings do not parrot words and actually understand them rather than memorizing spatial context. It’s like we are back at a stage when people thought we have finally developed AGI back when Goodfellows paper on GANs was published in 2014.

If you actually get off of the hype train u will realize most major industries use gradient boosting and achieve almost the same generalization performance for their needs as an LLM trained with giga fking tons of data. Because they can’t generalize well at all.

1