Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about memory consumption. #14

Open
Wxsxulin opened this issue Jul 5, 2024 · 4 comments
Open

Questions about memory consumption. #14

Wxsxulin opened this issue Jul 5, 2024 · 4 comments

Comments

@Wxsxulin
Copy link

Wxsxulin commented Jul 5, 2024

hank you for your work.

I have a few questions about your paper related to the memory consumption of the training.

1.What kind/How many GPUS did you use in your experiments?
2.How many GB of GPU memory do you use at each step?
3.May I ask you how much time in hours/days lasts the full training

Thanks in advance.

@dengxl0520
Copy link
Owner

  1. single 3090 GPU
  2. about 18G
  3. 100epochs took about 19 hours, CAMUS dataset.

@Wxsxulin
Copy link
Author

  1. 单个 3090 GPU
  2. 约18G
  3. 100epochs大约需要19个小时,CAMUS数据集。

thanks a lot

@jasonrichdarmawan
Copy link

My lab has 3090 GPU, but I cannot train the model

python train_video.py --modelname MemSAM --task CAMUS_Video_Semi
Namespace(modelname='MemSAM', encoder_input_size=256, low_image_size=256, task='CAMUS_Video_Semi', vit_name='vit_b', sam_ckpt='checkpoints/sam_vit_b_01ec64.pth', batch_size=1, n_gpu=1, epochs=100, base_lr=0.0001, warmup=False, warmup_period=250, keep_log=False, disable_memory=False, disable_reinforce=False, disable_point_prompt=False)
Single object mode: False
Total_params: 58047159
Segmentation fault

@xbbbbblv99
Copy link

Namespace(modelname='MemSAM', encoder_input_size=256, low_image_size=256, task='CAMUS_Video_Full', vit_name='vit_b', sam_ckpt='checkpoints/sam_vit_b_01ec64.pth', device='cuda', epoch=100, batch_size=2, n_gpu=1, base_lr=0.005, warmup=False, warmup_period=250, keep_log=False, frame_length=10, point_numbers=1, enable_memory=False, semi=False, reinforce=False, disable_point_prompt=False)

but

torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 318.00 MiB. GPU 0 has a total capacity of 23.68 GiB of which 266.69 MiB is free. Process 787029 has 23.42 GiB memory in use. Of the allocated memory 22.33 GiB is allocated by PyTorch, and 802.83 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants