ichiichisan
ichiichisan t1_j9k3sx1 wrote
Although this is interesting work, you are no lawyers and will not be able to provide "provable copyright protection".
ichiichisan OP t1_ivznfay wrote
Reply to comment by spurious_waffles in [D] Regularization & augmentation for NLP finetuning by ichiichisan
Thanks, but I am not looking for suggestions, but rather for something that has been proven to work, in best case with research on it.
It is quite common knowledge that any random altering of input text does not help with finetuning NLP tasks.
ichiichisan OP t1_ivza5fi wrote
Reply to comment by mediocregradstudent in [D] Regularization & augmentation for NLP finetuning by ichiichisan
In my current specific task it is simple multilabel regression, but I am also regularly working on other multilabel/multiclass classification and regression tasks.
And with low data sample I refer to ranges of 1k+ samples still, but I mostly work on longer text, not short sentences.
ichiichisan t1_jacz676 wrote
Reply to [D] More stable alternative to wandb? by not_particulary
Neptune.ai - I personally find wandb unbearable - also they collect by default all your training code which I find extremely shady.