Tensorflow OOM on GPU
I resolve this issue by reducing
Only to reduce memory use is to reduce batch_size.
Batch_size depends on your gpu graphics card, size of VRAM, Cache memory etc.
Please prefer this Another Stack Overflow Link
Try to take a look at this
Be careful not to run the evaluation and training binary on the same GPU or else you might run out of memory. Consider running the evaluation on a separate GPU if available or suspending the training binary while running the evaluation on the same GPU.