hellrail
hellrail t1_je2hg1z wrote
Reply to [D] I've got a Job offer but I'm scared by [deleted]
How about you offer to work for half of the payment they offered?
hellrail t1_jdtayym wrote
Reply to comment by iobeson in Why are humanoid robots so hard? by JayR_97
Yeah right, no one's investing in AI these days, because it doesnt work well enough
hellrail t1_jdsy919 wrote
Reply to comment by RastaNecromanca in Why are humanoid robots so hard? by JayR_97
There would be a great demand if one could build it, but we cannot yes due to limited AI
hellrail t1_jdsu8d9 wrote
Reply to comment by RastaNecromanca in Why are humanoid robots so hard? by JayR_97
Totally wrong
hellrail t1_j9wnk66 wrote
Wtf of course man, u also need one if you fit y=ax+b dude
hellrail t1_j5zov8n wrote
NEPTUNE
hellrail t1_j5y0ok5 wrote
Reply to comment by its_ya_boi_Santa in [D] Couldn't devs of major GPTs have added an invisible but detectable watermark in the models? by scarynut
Wrong, I am not in any rut.
New accounts, being in a rut, saying wrong just for the sake of saying something eventhough nothing was wrong....
If i look at your behaviour it clearly shows that you are fighting your own inner demons instead of really replying to what somebody has said (otherwise u wouldnt put so much of self-fantasized allegations in your posting).
I hope this kind of self-therapy works out for you, but i doubt it helps with anything.
hellrail t1_j5xwq72 wrote
Reply to comment by its_ya_boi_Santa in [D] Couldn't devs of major GPTs have added an invisible but detectable watermark in the models? by scarynut
Wrong. Its the very same account.
And in your previous answer, when u thought i was sb different, you already explained why you did it, now you claim to not remember. Hahaha. You are contradictory and nonsense as usual.
hellrail t1_j5xsq46 wrote
Reply to comment by its_ya_boi_Santa in [D] Couldn't devs of major GPTs have added an invisible but detectable watermark in the models? by scarynut
Ah, so there was nothing wrong in my statement but you just wanted to be obnoxious?
Good that you admit it.
PS: im the guy haha
hellrail t1_j5x5u8o wrote
Reply to comment by its_ya_boi_Santa in [D] Couldn't devs of major GPTs have added an invisible but detectable watermark in the models? by scarynut
And what exactly is wrong about the statement?
hellrail t1_itf78g0 wrote
Reply to comment by trutheality in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
Wrong.
Due to minimum 2 reasons:
- You dont take into account the time dimension. As doc Brown already said, "think 4-dimensional"!
It is a difference, if one has just misunderstood the theory, but could have known better. Or if that one has "done everything right" up to the point of science's knowledge, at that point of time.
In the first case, its the "users" fault, in the second, its on the current state of the theory. The first one is rubbish, only the second one yields knowledge. If you submit a paper of the first case, it gets rejected. If you submit one of the second one, you get attention.
You are just "projecting" everything into one pseudo-abstract (cp. Reason 2.) timeless dimension, washing away the core in my "theories of theories":
By misuse i would clearly say its of the first case category, where the "user" fucked it up, the second should NOT be labelled as misuse.
Reason for this partition as already said: the first category is rubbish, the second one is valueable.
Or another equally good reason: it cannot semantically be a misuse, if the "user" does not violate any assumption known at that time, because the current "rules" do define what a misuse is. Misuse is dynamic, not static. Think four dimensional, Doc Brown!
- It is further not true, that CS theories are purely abstract and basically formal logic and cannot be wrong.
Even in mathmatics itself, only a subset of the disciplines are purely abstract.
One example, where this is not the case, is coincidently the greatest mathmatical question currently: by what rule are the prime numbers distributed? This is in the field of number theory.
The earliest theory was, that they are log-distributed, by Gauß. His log approximation of the step function of primes was good, bur clearly deviates. Meanwhile, many other mathmatical theories, such as a log distribution, have been developed, leading up to the riemann hypothesis.
In this example u see, that a mathmatical model (here the log fct) is APPLIED TO DESCRIBE a natural phenomenon (the distribution of primes). Now the correctness is not decided in some formal level, but in the question if the natural phenomenon can be accurately enough described by the suggested mathmatical model.
What you say, that math. theories cannot be wrong because they are deduced/proven/whatever, is totally off-topic in this example, and could at maximum refer to the formal correctness of the theory around the log function involved here, e.g. the rules of adding two logs or similar, which are itself proven to be correct.
But that is not of interest here, the log theory is just taken, and suggested to model a natural phenomenon. The mathmatical theory here is that the step fct. of primes follows a log distribution.
By the way, there is also no conditions to be met, there is just one step fct. and one candidate theory to predict it.
The answer to the correctness of a theory about the distribution of primes does not lie in some formal deduction, but lies directly in the difference of the predicted and real distribution of primes.
In CS theories, of course you have questions from the CS domain that you want to answer with your mathmatical model. Only a small subset of CS theory deals with purely abstract proofs that stand for itself, e.g. in formal verification. The majority has a domain question to be answered and very often that question is quantitative, such that CS theories predict numbers that can be measured against real data measured from the phenomenon of interest, which then detetmines the correctness of the theory
hellrail t1_itewyaw wrote
Reply to comment by Real_Revenue_4741 in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
One thing i must add regarding the topic of presentation as "established knowledge".
The lecture you quoted, is lecture number 12. It is embedded in a course. There are of course lecture 11, 10, 9 etc. If you check these, which are also accessible with slightly midifying the given link, you see the context of this lecture. Specifically, a bunch of classifiers are explicitly introduced, and the v-dim theory on lecture 12 are still valid of these. The course does not adress deep networks yet.
So its a bit unfair to say these lecture does teach you a theory that deviates. Its does not deviate for the there introduced classifiers.
hellrail t1_iteuwcw wrote
Reply to comment by trutheality in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
Yes you are right. That is, because in most cases, not the theory is wrong but its wrongly understood, artifically extended, wrongly used.
It is therefore important to distinguish between deviations from theory, and deviations from misuse/misinterpretation.
hellrail t1_itdlgvb wrote
Reply to comment by Real_Revenue_4741 in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
Ok found the right one.
Well, generally i must say good example. I accepted it at least as a very interesting example to talk about, worth mentioning in this context.
Nevertheless, its still valid for all NON cnn, resnet, transformer models.
Taking into account, that its based on an old theory (prior 1990), where these deep networks have not existed yet, one might take into account its limitedness (as it doesnt try to model effects taking place during learning of such complex deep models, which hasnt been a topic back then).
So if I would be really mean, i would say u cant expect a theory making predictions about entities (in this case modern deep networks) that had not been invented yet. One could say that the v-dim theory's assumptions include the assumption of a "perfect" learning procedure (therefore exclude any dynamic effects from the learning procedure), which is still valid for decision trees, random forrest, svms, etc, which have their relevance for many problems.
But since im not that mean, i admit that this observations in these modern networks do undermine the practicability of the V-dimension view for modern deep networks of the mentioned types, and that must have been a mediocre surprise before having tried out if v-dims work for cnn/resnet/transformers, therefore good example.
hellrail t1_itde28m wrote
Reply to comment by Real_Revenue_4741 in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
Then please point mento the right slide by gibing the slide number
hellrail t1_itdde95 wrote
Reply to comment by Real_Revenue_4741 in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
- I disagree. Give me an example where the assumption is just a matter of semantics.
I state that every correct (and by that i mean scientific) formulation of assumptions can be even abstracted and formalized, and even incorporated in an automated algorithm yielding the answer weather this assumption is true or not, w.r.t the theories assumptions.
Proof: take an arbritrary assumption formulation and convert it to mathmatical formulation. Then us goesels numbers to formalize.
If you say now, well the conversion to a mathmatical formulation can be ambigious, i would ask you to clearly state the assumptions in a language that is suited for a scientific discussion.
- On the model selection slide, i see its just stated that model/hyperp optimization aims at selecting optimal parameters. Thats ofc trivially true.
If you Talk about the subsequent slides, i see it introduces one idea, to get some guidance in finding the opt settings, called bayesian occams razor. Occams razor is a HEURISTICS. Thats so to say the opposite of a rule/theory.
A property of heuristics is explicitly, that it does not guarantee to yield a true or optimal solution. A heuristics can by definition not be wrong or correct. its a heuristics, a strategy that has worked for many ppl in the past and might fail in many cases. A heuristics does not claim to provide a found rule or similiar.
Now on the last slide they even address the drawbacks of this heuristics. What do you expect more?
As i expected, this is not an example of a theory stating something that deviates from reality. Its just a HEURISTIC strategy they give you at hand, when you want to start with hyperparameter finding but you have no clue how. Thats when you go back to heuristics (please wikipedia heuristics) and i bet this proposed heuristics is not the worst you can do even today, where more knowledge has been accquired.
hellrail t1_itd4jqh wrote
Reply to comment by Real_Revenue_4741 in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
- It is not a matter of semantics. Your interpretation included all reasons for deviations, that are caused by a "wrong usage", for which i have made an extrem example. My interpretation does exclude these cases. Distinguishing between wrong and correct usage can be clearly pinned to the assumptions of the experiment. To pick up my example from before: a guy trying to predict fluid dynamics with maxwells eqs started with the assumptions that maxwells eqs are capable of predicting fluid dynmaics. That is objectively wrong. Especially, it CANNOT be interpreted as "his assumption is correct, maxwells eqs just lack of certain aspects necessary to describe fluid dynamics". No.
The wrong usage of a theory can be pinned down to the assumptions, and these are not a matter of semantics.
- Not knowing the class you have visited, i can only comment on this if you link me that lecture, such that i can see what really was thought there as established theory and what not.
In classes i have visited or seen in the internet, i have never seen somebody stating that its a global rule that larger models, without exception, do increase the danger of overfitting or similar. Such topics were discussed at maximum in context of "intuition", resp. The teacher just shared his own experiences. And still, thats often true.
But i am open to see the example lecture, that teaches that as a general rule explicitly, such that it has been falsified later.
hellrail t1_itczie0 wrote
Reply to comment by 0ffcode in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
Not a well established theory, just an implemented software feature being advertised here.
And its trivially true, that when a method that tries out more hyperparameters and more models, the probability increases that it spits out a configuration that works well for you (the only statement in this that comes close to a "theory").
Weather an implemented software is doing that up to a satisfactory manner for a user, or if it can be achieved at all, is a question in the domain of engineering, not theory.
hellrail t1_itcw9n1 wrote
Reply to comment by Real_Revenue_4741 in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
@ "When practice deviates from theory, this usually means that the theory does not well-capture the results that people are getting in practice. This does not necessarily mean that the theory is incorrect, but that usually the implications of the theory and it’s common inductions don’t capture the entire picture."
Thats a rather stupid (sorry) interpretation of what the term "deviations from the theory" should mean. If one follows your interpretation, it means that you would declare an experiment, where someone tries to calculate some fluid dynamics by using maxwells equations and then gets results that not match the measurements, as a "deviation from the theory".
Thats nonsense, because obviously the theory IS JUST USED wrong: the theory around the Maxwell equations NEVER suggested to model the dynamics of fluids.
therefore, a more sensical interpretation for a "theory deviating from reality" is to include the assumption, that the theory "is used correctly" (meaning theory predictions are compared with measurements, for which the theory really is made for to predict up to a certain accuracy).
If then the theory is applied correctly, but the measurements deviate qualitatvely from the measured reality, that implies that the theory does not model the real mechanics accurately and therefore is wrong (in science we call that falsification).
At the example you gave: you named some ongoing discussion in an open research field. Of course, when humans try to come up with explanations for something (when they develop NEW theories), 99% of the ideas are wrong at first and are still being discussed. Through experiments and falsification, theories are adapted until they are not falsified anymore. That means, every theory, with time, gets filtered by the scientific process, until more and more experiments confirm it by not falsifying it. At the end of this process, a theory is well established and starts to being taught e.g. to students in universities.
Your example is not one for a well established theory, that the topic creator as a student could have learned as "truth" about in one of his lectures. And the topic name is "what .... did you LEARN".
hellrail t1_itcmak6 wrote
Reply to comment by mediocregradstudent in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
Does that mean that MLP are not universal function approximators? No.
Its a fact that MLP is capable of fitting arbritrary functions.
Does anything here deviate from the theory? No.
hellrail t1_itclbpy wrote
Reply to comment by 4bedoe in [D] What things did you learn in ML theory that are, in practice, different? by 4bedoe
As i see you lack of examples 😄
hellrail t1_itcksv1 wrote
None, also not in other disciplines
hellrail t1_isuiy2x wrote
There would be a lot more explosions happening blowing up the whole house
hellrail t1_iskhso6 wrote
Reply to comment by kakhaev in [P] I built densify, a data augmentation and visualization tool for point clouds by jsonathan
@ Usually augmentation allow you to increase sample of your input/output space that will lead to better map function that your model will learn.
More data better results in general yes, but if the additional data is worthless, its a bit scam. That will be recognized in a comparison with an equally well trained state without that augmentation (might be harder to reach) tested on relevant data.
Technically put: the learned distribution is altered to a surrogate pointcloud which is quite similar to the relevant distribution of sensor data that will be produced measuring the real world, but is not the same anymore. Thats the price for more training data with this, and i wouldnt pay it because my primary goal is to capture the relevant distribution as Close as possible.
hellrail t1_je51otz wrote
Reply to What science and technology should be here already (2023) but isn’t? by InfinityScientist
Nothing, because what not is that shall not be yet, dummy