Submitted by ichiichisan t3_ys974h in MachineLearning
I am trying to research well working methods for regulraization in small-data NLP finetuning scenarios, specifically for regression.
Coming from computer vision background, it appears to me that no established method has emerged that works well across tasks and it is really hard to combat stark overfitting on small data tasks.
I am specifically looking for methods that are special to NLP finetuning and go beyond classical DL regularization techniques like dropout or weight decay.
Happy for any pointers!
mediocregradstudent t1_ivyynyr wrote
Recent work has shown generating a paraphrase of the original sentence could be used to improve robustness for sentence-level NLP tasks. What specific task are you working on in the low data setting?