Viewing a single comment thread. View all comments

DSM-6 t1_javnmz2 wrote

Personally, I think the answer is existing bias in the training data.

I don’t know enough about chatgpt to state this as fact, but I think it’s safe to assume that chatgpt understands or adheres to grammar rules. I.e. nowhere in the code does it state “antecedent pronouns should refer to the subject of a sentence”

Instead I assume chatgpt grammar comes repeated convention in the training data. Enough data in which the antecedent refers to something other then the sentence object means that the “they” can refer to any of the preceding nouns. In that case “councilmen fear voilence” is a far more common sentence in the training than “protesters fear violence”

Then again your example was passive tense, so I dunno 🤷‍♀️.

5

New_Computer3619 OP t1_javqiuu wrote

I tried the same questions in separate chats as in the edited post. ChatGPT gave incorrect/unsatisfying answers this time. May be without context from previous Q&A, it can only infer using grammar rule? What do you think?

2