Submitted by natural_language_guy t3_ypxyud in MachineLearning
I am trying to build an NER model, but want multiple options for the spans...ex:
"I like green cats." -> {BOBI, BIII, BOOO, etc}
that I can feed into another algorithm to choose based on downstream criteria.
​
With something like T5, I would modify the beam search to give me a list of generative texts from most probable to nth most probable. With BERT, I don't know how to do this because I can't condition the result of a token classification on the previous one.
fastglow t1_ivlif23 wrote
Have you looked into nested NER? Various approaches exist, e.g. span-based, hypergraph-based, etc. Check this paper for example: https://arxiv.org/abs/2203.10545