Comments

You must log in or register to comment.

izumi3682 OP t1_ixk6ojd wrote

Submission statement from OP. Note: This submission statement "locks in" after about 30 minutes, and can no longer be edited. Please refer to my statement they link, which I can continue to edit. I often edit my submission statement, sometimes for the next few days if needs must. There is often required additional grammatical editing and additional added detail.


Here is the research paper.

https://www.science.org/doi/10.1126/science.ade9097

From the article.

>To create Cicero, Meta pulled together AI models for strategic reasoning (similar to AlphaGo) and natural language processing (similar to GPT-3) and rolled them into one agent. During each game, Cicero looks at the state of the game board and the conversation history and predicts how other players will act. It crafts a plan that it executes through a language model that can generate human-like dialogue, allowing it to coordinate with other players.

>Meta calls Cicero's natural language skills a "controllable dialogue model," which is where the heart of Cicero's personality lies. Like GPT-3, Cicero pulls from a large corpus of Internet text scraped from the web. "To build a controllable dialogue model, we started with a 2.7 billion parameter BART-like language model pre-trained on text from the Internet and fine tuned on over 40,000 human games on webDiplomacy.net," writes Meta.

>The resulting model mastered the intricacies of a complex game. "Cicero can deduce, for example, that later in the game it will need the support of one particular player," says Meta, "and then craft a strategy to win that person’s favor—and even recognize the risks and opportunities that that player sees from their particular point of view."

So, my question is, is this an "incremental improvement" in our AI development efforts, or is this more like the "AI significantly improves every three months" level of improvement.

https://www.ml-science.com/exponential-growth

Are we seeing any evidence that AI of any form is improving significantly every 3 months?

5

AcademicGuest t1_ixk6r37 wrote

Nope, it’s all invalid because it is not performed by a human.

−11

izumi3682 OP t1_ixk7qay wrote

What is not performed by a human? The game play? I thought the point was that the AI was learning to outplay humans in highly sophisticated incomplete information games. If I am misunderstanding your point, please explain to me what you mean.

5

jwg020 t1_ixk84dp wrote

I mean, I feel like AI would be better at ruling strictly on logical reason as opposed to religious fanaticism, corporate greed, etc. I’m all for it.

5

Coachtzu t1_ixk8o3i wrote

Strict logic doesn't always work in the real world though. Sometimes you need an empathetic voice in the room. We have plenty of faults governing ourselves, but I'm not sure the AI should be trusted to find answers that aren't necessarily for the greater good, but benefit those not in charge.

4

izumi3682 OP t1_ixkb038 wrote

>AI represents, unless strictly curtailed and linear, an inherit violation of human will and freedom

I am still not sure what this has to do with the development of ever more powerful AI technology. But ok.

6

FuturologyBot t1_ixkdqvd wrote

The following submission statement was provided by /u/izumi3682:


Submission statement from OP. Note: This submission statement "locks in" after about 30 minutes, and can no longer be edited. Please refer to my statement they link, which I can continue to edit. I often edit my submission statement, sometimes for the next few days if needs must. There is often required additional grammatical editing and additional added detail.


Here is the research paper.

https://www.science.org/doi/10.1126/science.ade9097

From the article.

>To create Cicero, Meta pulled together AI models for strategic reasoning (similar to AlphaGo) and natural language processing (similar to GPT-3) and rolled them into one agent. During each game, Cicero looks at the state of the game board and the conversation history and predicts how other players will act. It crafts a plan that it executes through a language model that can generate human-like dialogue, allowing it to coordinate with other players.

>Meta calls Cicero's natural language skills a "controllable dialogue model," which is where the heart of Cicero's personality lies. Like GPT-3, Cicero pulls from a large corpus of Internet text scraped from the web. "To build a controllable dialogue model, we started with a 2.7 billion parameter BART-like language model pre-trained on text from the Internet and fine tuned on over 40,000 human games on webDiplomacy.net," writes Meta.

>The resulting model mastered the intricacies of a complex game. "Cicero can deduce, for example, that later in the game it will need the support of one particular player," says Meta, "and then craft a strategy to win that person’s favor—and even recognize the risks and opportunities that that player sees from their particular point of view."

So, my question is, is this an "incremental improvement" in our AI development efforts, or is this more like the "AI significantly improves every three months" level of improvement.

https://www.ml-science.com/exponential-growth

Are we seeing any evidence that AI of any form is improving significantly every 3 months?


Please reply to OP's comment here: https://old.reddit.com/r/Futurology/comments/z36el2/training_our_future_rulers_meta_researchers/ixk6ojd/

1

mapadofu t1_ixkerit wrote

Dude, I could never master Diplomacy despite playing many times.

1

Zacpod t1_ixkmlqu wrote

I, for one, welcome our AI overlords. They can't possibly be worse than the power hungry sociopaths we vote in.

20

ABrokenBinding t1_ixl2j5l wrote

Somehow a tool that can trick humans with natural language, placed in the hands of DJ Markie Z, doesn't seem like something's scholars would call "good".

Just me?

4

Swordbears t1_ixl5yc0 wrote

The AIs will be owned and controlled by the wealthy if we don't fix this shit first. Our AI overlords are most likely going to be better at oppressing and exploiting us for the sake of the few.

5

Moonbase0 t1_ixlttj1 wrote

Is this Night Mother related? Because all I can hear is "Let's kill someone"

1

Sidoplanka t1_ixluzp5 wrote

Using "Meta", "future rulers" and "diplomacy" in the same sentence is just beyond silly 😂

3

Nyarlathotep854 t1_ixlxnh3 wrote

Honestly, as stupid as this narrow application sounds, i am excited for what this means for strategy games

1

Kaionacho t1_ixlyybc wrote

As long as the AI doesn't accept bribes that would already be better then what we have now

1

hungrycryptohippo t1_ixma79t wrote

So isn’t the title of this post misleading? The AI trained on existing data and used simulators so it’s only good at learning how to play diplomacy specifically and would need more data to be applied to any other domain.

Still an impressive result but geez, it’s not like we have a general system here that can be interfaced to a new problem and do well, especially if there isn’t data like there was for diplomacy from other humans.

1