Submitted by floppy_llama t3_1266d02 in MachineLearning
Swolnerman t1_jead4wo wrote
Reply to comment by drizel in [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
How can it do that with a context window of 32k?
On top of that, I don’t think gpt4 can make informed decisions on picking between academic research papers as of yet
Viewing a single comment thread. View all comments