-
Notifications
You must be signed in to change notification settings - Fork 82
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error running code cnn-train code chunk in 02-cats-vs-dogs.Rmd #10
Comments
Just adding a note that this issue is related to running code on Rstudio Server (including GPU). |
I can reproduce this problem. Investigating. I think I'm running out of GPU memory. Which wasn't a problem before. I do have two sessions running but not sure if that's relevant. |
Dropping the batch size to 5 has got it moving again. |
I've tried dropping the batch size to 5, but am still getting errors. The code progresses through all 20 epochs, whereas it was stopping at the first with a larger batch size
|
When running the following code chunk from a fresh session I get the following error:
Error in py_call_impl(callable, dots$args, dots$keywords) : ResourceExhaustedError: OOM when allocating tensor with shape[6272,512] and type float on /job:localhost/replica:0/task:0/device:GPU:0 by allocator GPU_0_bfc [[node MatMul_3 (defined at /util/deprecation.py:324) ]] Hint: If you want to see a list of allocated tensors when OOM happens, add report_tensor_allocations_upon_oom to RunOptions for current allocation info. [Op:__inference_distributed_function_1290] Function call stack: distributed_function
The text was updated successfully, but these errors were encountered: