Submitted by inFamous_16 t3_11hq1ga in deeplearning
I have tweets and the task is to perform text classification. I already have learned token embeddings for those tokens present in each of the tweets through some Graph based NN model. Now that I want to use those token embeddings to represent the tweet but the issue is every tweet will have different size embeddings if I just do concatenation. Is there any way, where I can input variable length embeddings to pre-trained BioBERT (if not, any other BERT) model and still be able to perform classification task?
I_will_delete_myself t1_jauuhhi wrote
You add padding