Submitted by NLP2829 t3_yu8nna in MachineLearning
(I only want to do inference, I don't need to finetune it.)
I want to use very-large language model (#parameters > 100B) to do some experiments, is that true the only very-large language model we can get access to is GPT3 API? Can we possibly get access to PaLM and Flan-PaLM 540B with no cost by chance?
I have searched over the internet but can't find a definite answer. As GPT-3 pricing for text-davinci-2 is not cheap, I am wondering if there's a chance to use other models.
Also, I can request up to 372GB VRAM, is there any large language model (#parameters > 100B) that I can actually download and run "locally"?
allwordsaremadeup t1_iw83d9j wrote
Bloom not big enough for you? 176B parameters. Can be downloaded here: https://huggingface.co/bigscience/bloom