Submitted by Haghiri75 t3_11wdi8m in deeplearning
Recently, I installed dalai on my Macbook Pro (late 2019, i7 processor and 16GB of RAM) and I also installed Alpaca-7B model. Now when I ask it to write a tweet, it writes a wikipedia article and it does the same pretty much every time 😂
First, should I fine-tune it?
Second, is there any "prompt magic" going on here?
P.S: using this one, I got much better results. What's the difference between the two?
Haghiri75 OP t1_jcxt80d wrote
I guess I found the reason. Dalai system does quantization on the models and it makes them incredibly fast, but the cost of this quantization is less coherency.