Nextil t1_jb1sg1c wrote on March 5, 2023 at 7:54 PM Reply to comment by Art10001 in [R] RWKV (100% RNN) can genuinely model ctx4k+ documents in Pile, and RWKV model+inference+generation in 150 lines of Python by bo_peng I think they mean with offloading/streaming you need 3GB minimum, but it's much slower. Permalink Parent 1
Nextil t1_j1zqxp9 wrote on December 28, 2022 at 4:47 PM Reply to comment by EthansWay007 in [D] When chatGPT stops being free: Run SOTA LLM in cloud by _underlines_ You can rate the responses up or down and provide an "ideal" response. Permalink Parent 2
Nextil t1_jb1sg1c wrote
Reply to comment by Art10001 in [R] RWKV (100% RNN) can genuinely model ctx4k+ documents in Pile, and RWKV model+inference+generation in 150 lines of Python by bo_peng
I think they mean with offloading/streaming you need 3GB minimum, but it's much slower.