I plan to launch a very large recurrent network (for example, 2048x5), is it possible to define one layer on one GPU in a tensor stream? How should I implement the model for maximum efficiency. I understand that there is overhead for exchanging data between the GPU or GPU-CPU-GPU.
python neural-network tensorflow
read Read
source share