xirzon t1_itfkplo wrote
Reply to comment by ChronoPsyche in Given the exponential rate of improvement to prompt based image/video generation, in how many years do you think we'll see entire movies generated from a prompt? by yea_okay_dude
The paper "Re3: Generating Longer Stories With Recursive Reprompting and Revision" shows some interesting strategies to work around that limitation by imitating aspects of a systematic human writing process to keep a story consistent, detect errors, etc.: https://arxiv.org/abs/2210.06774
A similar approach is taken by the Dramatron system to create screenplays and theatre scripts: https://arxiv.org/abs/2210.06774
In combination with more systematic improvements to LLM architecture you hint at and next-gen models, we might see coherent storytelling sooner than expected (with perhaps full length graphic novels as the first visual artform).
ChronoPsyche t1_itflq78 wrote
Oh there are certainly workarounds! I agree 100%. These workarounds are just that though, workarounds. We won't be able to leverage the full power of long-form content generation until we solve the memory issues.
Which is fine. There is still so many more advances that can be made in the space of the current limitations we have.
visarga t1_itgqug0 wrote
There is also exponentially less long-form content than short form. The longer it gets, the fewer samples we have to train on.
Viewing a single comment thread. View all comments