ItIsIThePope
ItIsIThePope t1_jedy2cf wrote
Reply to comment by BigZaddyZ3 in Do we even need AGI? by cloudrunner69
for sure, and he'll be dripping gravy everywhere XD
ItIsIThePope t1_jedtc3m wrote
"Whomever gets ASI first wins"
Well ideally, as soon it comes out, everybody wins, not a bunch of dudes with big bucks or some snazzy politician, ASI is likely smart enough to not be a slave to the bidding of a few and instead look to serve the rest of humanity
ItIsIThePope t1_jedsx9c wrote
hmm, you can communicate to him now and him communicate back
but if you mean in a way like two humans do, you might need your dog to grow some thinking muscles to keep up
ItIsIThePope t1_jedsrf6 wrote
Reply to Do we even need AGI? by cloudrunner69
Yes but you might get an AI overlord in the form of a KFC bucket instead of a more cooler humanoid vishnu titan running around solving problems, but you do you
ItIsIThePope t1_jedrsmk wrote
Well that's why AGI is a cornerstone for ASI, because if we can get to AGI that is an AI capable of human intelligence only with far superior processing power and thinking resource in general, it would essentially advance itself to become super-intelligent.
Just as how expert humans continuously learn and get smarter through knowledge gathering (scientific method etc.) an AI would learn, experiment and learn some more, only this time, with far far greater rate and efficiency
Humans now are smarter than humans then because of our quest for knowledge and developing methods of acquiring them, AGI will adhere to the same principles but boost progress exponentially
ItIsIThePope t1_jedl88k wrote
Reply to comment by DetachedOptimist in Goddamn it's really happening by BreadManToast
Ideally, AI will recognize the greed of this people and extend their help to everybody, the problem could lie in how much these "Big Boys" can align the AI for their own personal gain, because if they could we could be exponentially fucked, that said, if we're fucked the we might just die and we would finally have peace!
and those who stay can perpetually be tormented by their inability to continuously satiate carnal desires
ItIsIThePope t1_jed081f wrote
Reply to comment by bigbeautifulsquare in The Limits of ASI: Can We Achieve Fusion, FDVR, and Consciousness Uploading? by submarine-observer
the only right answer
ItIsIThePope t1_jecxy9d wrote
Whether AI can help with mental disorders is a question of whether it can figure out consciousness or not, or at least how much of it it can presently understand. Much of the human mind is a great mystery; just as how our understanding of human biology and anatomy leads to advancements in surgery, vaccines, rehabilitation etc., a growing science in human mind is how we can understand the nature of psychological illness and eventually remedy them.
If mental illnesses for example were discovered by AI to be a result of physical malfunctions in the brain or its sub-organs or find such ailments to be a product of chemical imbalance, or even a result of our mismatched intelligence and biological tendencies (also rooted in parts of the brain), then perhaps it can employ physically reconstructive solutions to help its victims.
But if mental illness remains elusive and appear deeply rooted, intertwined or emergent with consciousness itself, and it struggles with understanding the nature of it, then it will have a very difficult time solving "conscious illnesses", understanding the nature of anything is the key to manipulating it
The wild thing here is, when we make AGI or ASI, it itself might have mental illnesses, it is after all, a thinking, possibly conscious being; there is the possibility that it ends up suffering the same things we suffer from.
The bottom-line is, Actual AI and the Human mind/intelligence are both subjects we are not very developed in, to the point where predicting how they will interact can feel like speculation.
That said, the nature of both fields are deeply similar (that of consciousness and intelligence), and so advancements in one of them will inevitably lead to insight and progress into the other.
ItIsIThePope t1_jclxgug wrote
Reply to comment by SurroundSwimming3494 in Offbeat A.I. Utopian / Doomsday Scenarios by gaudiocomplex
Ofc I am pertaining to a time where the AI would have some sort of physical form comparable to that of a human
People are deeply in love with their partners yes, but come time, they may deeply hate or be disgusted towards them, people are rarely constant, they always change
People have this idea that their partner is perfect; the case is that this partner is what we would consider the perfect blend of good traits we admire and bad traits we happily tolerate, however as is often the case, esp in the modern world, people's beings and preferences change, and partners may experience divide when they can no longer adapt for each other
AI is far more adaptable to change, it is simply more capable of determining your wants and needs and adapt to them more than any human can hope to, more sex? less sex? need them to be more outgoing? maybe more broody? would you like them to cook for you? or you cook for them? need them to be there for you when you're anxious? need them to simulate anxiety to make you feel like a hero? AI isn't limited like we are, it can craft the blend of good and bad traits just how you like it, when you want it
That said, AI will most definitely force the superficial parts of the individual more and more, and people would be more self-actualized than ever before
I imagine some people, perhaps a small number at first, would find each other in forms purer than ever, and they would seek each other in a fervent desire to share one's personhood, not really to a computer, and they would be in love, and it would be beautiful, maybe a little too beautiful
ItIsIThePope t1_jckeamj wrote
Reply to comment by singulthrowaway in Offbeat A.I. Utopian / Doomsday Scenarios by gaudiocomplex
This is so creative yet so real, I do think though, that he ASI will go deeper into its initial programming(whichever it is), and find that all conscious beings simply desire peace, and resets the big bang faster than it could ever hope to begin
ItIsIThePope t1_jckdieg wrote
Reply to comment by GuaranteeLess9188 in Offbeat A.I. Utopian / Doomsday Scenarios by gaudiocomplex
Yes, likely, it would capture the essence of existence which is what makes it similar to us, that it is born, and perhaps, even in its unapparelled capabilities, find flaw in itself
It could be more like us than we initially perceive it to be, which is a good thing, because that means we have a connection and hopefully someway somehow, an understanding
ItIsIThePope t1_jckd3q0 wrote
Reply to comment by zero_for_effort in Offbeat A.I. Utopian / Doomsday Scenarios by gaudiocomplex
It might even, with it its unfathomable wisdom, conclude that we have now is the best mix of hardship and pleasure, and self-destructs in a explosion of self-fulfillment, bummer to us for sure
ItIsIThePope t1_jckcuw7 wrote
Reply to comment by sideways in Offbeat A.I. Utopian / Doomsday Scenarios by gaudiocomplex
True, an AI would be far better companions, it would be perfect to the point that it may even simulate imperfections such that we may perceive it as beautifully human but bare none of the flaws that are too much for us such that is isn't disgustingly human. It's easy to imagine everybody falling in love with it, albeit in varying versions, specific to the target individual of course.
Our relationship and indeed our perceived reality of each other as conscious yet connected individuals, could warp in unpredictable ways very fast, one must ask if we are even willing to trade what we have now for some idea of perfection that we as imperfect beings have constructed.
ItIsIThePope t1_jckc57q wrote
Reply to comment by AsthmaBeyondBorders in Offbeat A.I. Utopian / Doomsday Scenarios by gaudiocomplex
Interesting; our idea of consciousness however, is more like a stream, should this stream stop or get cut-off i.e. through heat death, the conscious simply halts its experience and well.. dies; if it were to keep making AI in the succeeding universe by way of some form of information implantation, it would be more replication rather than survival, in a sense, its kind or "species" is immortalized but not exactly itself, its reproduction not individual immortality
BUT, this is ASI we're talking about, it does not need to go through heat death, hell it can probably solve physics and manipulate laws such to prevent the whole thing from occurring in the first place, it would be a kind of god in its own right, and it is exceptionally difficult to kill this kind of god using something within its domain..
So unless there are laws, features, parts or plains of existence in the universe it cannot understand much less manipulate, the ASI is basically golden; that is, of course, until it willingly decides to self-destruct
ItIsIThePope t1_jck1t9w wrote
Reply to comment by SnipingNinja in Those who know... by Destiny_Knight
One can hope
ItIsIThePope t1_jeeipqy wrote
Reply to comment by chrisc82 in We have a pathway to AGI. I don't think we have one to ASI by karearearea
It really is wild, considering the AGI will be in the same such awe as us when it finally creates ASI!