Submitted by austintackaberry t3_120usfk in MachineLearning
Daveboi7 t1_jdndvq0 wrote
Reply to comment by dreamingleo12 in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
With databricks?
dreamingleo12 t1_jdndzmt wrote
No I don’t use databricks. I only tried LLaMA and Alpaca.
Daveboi7 t1_jdnedrd wrote
But which cloud service did you use to train them?
I tried using databricks to train a model but the setup was too complicated.
I’m wondering is there a more straightforward platform to train on?
dreamingleo12 t1_jdnel6b wrote
You can just follow Stanford Alpaca’s github instructions, as long as you have LLaMA weights. It’s straightforward.
Daveboi7 t1_jdneqdx wrote
Ah. I’m trying to train the Dolly model created developed databricks.
dreamingleo12 t1_jdnewt2 wrote
It’s just Alpaca with a different base model. Databricks boasted too much.
Daveboi7 t1_jdnf18o wrote
Yeah but the comparisons I have seen between Dolly and Alpaca look totally different.
Somehow the Dolly answers look much better imo
Edit: spelling
dreamingleo12 t1_jdnf4qn wrote
I don’t trust DB’s results tbh. LLaMA is a better model than GPT-J.
Daveboi7 t1_jdnf96e wrote
Somebody posted results on Twitter, they looked pretty good. I don’t think he worked for DB either. But who knows really
Viewing a single comment thread. View all comments