Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Time and Space Limits for Deep Learning on UoS HPCs #18

Open
TheLostLambda opened this issue Feb 18, 2023 · 1 comment
Open

Time and Space Limits for Deep Learning on UoS HPCs #18

TheLostLambda opened this issue Feb 18, 2023 · 1 comment

Comments

@TheLostLambda
Copy link

Hello! I'm looking to retrain / train a larger model inspired by https://github.com/nadavbra/protein_bert and was wondering:

  1. Where should I store training data? It's >1TB but it seems most storage locations that large on the Sheffield HPCs are temporary storage only?
  2. Could that data be persisted and could I reserve GPU usage for a few weeks to a month? The original model was trained for a month, but this was likely using a lower-powered GPU!

The GitHub for ProteinBERT seems to have some pretty nice instructions for the retraining, I'm just wondering if I'll run into any issues using that much disk space and that much GPU time!

Thanks a ton,
Brooks

@willfurnass
Copy link

I suggest the /fastdata (Lustre filesystem) areas as they don't have quotas and they're the most performant areas when working with larger files. Files in those areas will be deleted after 60 days - is that too little time?

Can the training process easily be paused/resumed? If so, that gives you more flexibility as to how the workload can be run: it could potentially be split across multiple jobs, possibly using free CPU/GPU cycles (see https://docs.hpc.shef.ac.uk/en/latest/hpc/scheduler/index.html#preemptable-jobs)

If you want to discuss your GPU resource needs in more detail then do get in touch via [email protected].

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants