Submitted by MohamedRashad t3_y14lvd in MachineLearning
_Arsenie_Boca_ t1_irwat46 wrote
Reply to comment by MohamedRashad in [D] Reversing Image-to-text models to get the prompt by MohamedRashad
Thats a fair point. You would have a fixed length for the prompt.
Not sure if this makes sense but you could use an LSTM with arbitrary constant input to generate a variable-length sequence of embeddings and optimize the LSTM rather than the embeddings directly.
Viewing a single comment thread. View all comments