Viewing a single comment thread. View all comments

ollih12 t1_izre2kw wrote

What is the best approach for text generation?

For context: I'm trying generate episode synopsis of a show by training a model with existing episode titles and synopsis of a show and using an input title as the input for the generated episode. I've read that LSTM models are good for this since they maintain the context. I have also read that GPT-3 is the best for this but it's not free. This is just a personal project and I intend on using PyTorch if it's of any significance, currently I have scraped synopsis and titles of existing episodes and have them stored in a pandas dataframe so just not sure where to go from here.

1

pythoslabs t1_izxk8im wrote

>also read that GPT-3 is the best for this but it's not free.

Try ChatGPT (https://chat.openai.com/ ) . Its free pre-beta release and so you can try your hands on it .

Also be careful that it might not be 100% factually accurate . But to try out simple text generation, it should do the job pretty well.

2

ollih12 t1_izzhdj9 wrote

Can ChatGPT be fine tuned for what I described?

1

BrightCounter738 t1_izzwtbu wrote

It is not open-sourced (and one probably wouldn’t be able to run it personally even if it was), so no.

1

ollih12 t1_j00bp15 wrote

Would the GPT-2 model from the transformers package be ok for it?

1

pythoslabs t1_j00ffog wrote

Yes.

You have to train on their system with your custom data. It is costly though.

eg: if you want to train on the Davinci model will cost you - $0.0300 / 1K tokens for training ( fine tuning ) and $0.1200 / 1K tokens for its usage - if you wish to use it as an API end point )

1

ollih12 t1_j00hiav wrote

Are there any free alternatives you would recommend?

1