GPU Memory Requirements for Training Large-scale se_atten Models #3902
yanjin-tech
started this conversation in
General
Replies: 1 comment
-
There are some benchmark results in our published paper. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello,
I would like to inquire about the pre-trained model used in the development of se_atten. Specifically, on what GPU with how much memory was it trained?
Recently, I have been working on a model with a very large dataset. Due to the variety in structural elements, space groups, and the number of atoms, the memory required for training data is significantly larger compared to training a model with a fixed structure. I often encounter memory error issues, and therefore, I am wondering if my problem is due to insufficient computational resources.
Beta Was this translation helpful? Give feedback.
All reactions