Hi, I was trying the seq2seq tutorial of tensorflow. It has 20 GB dataset. My VM on Google cloud compute is throwing me out of memory error while unpacking the data file. I have 8vCPUs, 10 GB instance. But that's ram, and I have 30GB disk. How do you guys do experiments with such data? Using AWS EMR ?