-
Notifications
You must be signed in to change notification settings - Fork 36
Description
Dear HamGNN Development Team,
I'm currently working on fine-tuning the HamGNN model for band structure prediction using silicon 3×3×3 supercells. Despite implementing several memory optimization strategies, I'm experiencing GPU memory overflow during the fine-tuning phase.
Here's my current setup and what I've tried:
Task: Fine-tuning for band structure prediction
System: Si 3×3×3 supercells
Batch size: Reduced to 1
Memory status: ~27.9GB/32GB used before fine-tuning
The memory overflow occurs specifically during the fine-tuning stage, even with these optimizations in place.
I would appreciate your insights on:
What could be causing this memory overflow despite the minimal batch size?
Are there specific memory optimization techniques you recommend for fine-tuning with large supercells?
What GPU models and memory configurations does your team typically use for training with 3×3×3 supercells?
Thank you for your time and assistance. I'm happy to provide any additional details about my setup if needed.
Best regards