somebodyenjoy OP t1_iycyur2 wrote
Reply to comment by HiPattern in If the dataset is too big to fit into your RAM, but you still wish to train, how do you do it? by somebodyenjoy
I do the same using numpy files, but they only let me load the whole data which is too big in the first place. Tensorflow let’s us load in batches huh, I’ll look into this
HiPattern t1_iyd91t0 wrote
hdf5 files are quite nice for that. You can write your X / y datasets in chunks into the file. When you access a batch, then it will only read the part of the hdf5 file where the batch is.
​
You can also use multiple numpy files, e.g. one for each batch, and then handle the file management in the sequence generator.
somebodyenjoy OP t1_iydqmu0 wrote
This is perfect, I won’t have to invest in additional RAM. Thanks for the tip!
Viewing a single comment thread. View all comments