[N] OpenAI may have benchmarked GPT-4’s coding ability on it’s own training data Submitted by Balance- t3_124eyso on March 28, 2023 at 5:57 AM in MachineLearning 138 comments 925
jabowery t1_je107nj wrote on March 28, 2023 at 5:09 PM See these entries in the Hutter Prize FAQ: Why aren't cross-validation or train/test-set used for evaluation? Why is (sequential) compression superior to other learning paradigms? Why is Compressor Length superior to other Regularizations? Why not use Perplexity, as most big language models do? Is Ockham's razor and hence compression sufficient for AGI? Permalink 6
Viewing a single comment thread. View all comments