Submitted by AutoModerator t3_zcdcoo in MachineLearning
ollih12 t1_izre2kw wrote
What is the best approach for text generation?
For context: I'm trying generate episode synopsis of a show by training a model with existing episode titles and synopsis of a show and using an input title as the input for the generated episode. I've read that LSTM models are good for this since they maintain the context. I have also read that GPT-3 is the best for this but it's not free. This is just a personal project and I intend on using PyTorch if it's of any significance, currently I have scraped synopsis and titles of existing episodes and have them stored in a pandas dataframe so just not sure where to go from here.
pythoslabs t1_izxk8im wrote
>also read that GPT-3 is the best for this but it's not free.
Try ChatGPT (https://chat.openai.com/ ) . Its free pre-beta release and so you can try your hands on it .
Also be careful that it might not be 100% factually accurate . But to try out simple text generation, it should do the job pretty well.
ollih12 t1_izzhdj9 wrote
Can ChatGPT be fine tuned for what I described?
BrightCounter738 t1_izzwtbu wrote
It is not open-sourced (and one probably wouldn’t be able to run it personally even if it was), so no.
ollih12 t1_j00bp15 wrote
Would the GPT-2 model from the transformers package be ok for it?
pythoslabs t1_j00ffog wrote
Yes.
You have to train on their system with your custom data. It is costly though.
eg: if you want to train on the Davinci model will cost you - $0.0300 / 1K tokens for training ( fine tuning ) and $0.1200 / 1K tokens for its usage - if you wish to use it as an API end point )
ollih12 t1_j00hiav wrote
Are there any free alternatives you would recommend?
Viewing a single comment thread. View all comments