How to build CUDA JIT caches for all available kernels in TensorFlow programmatically?
问题 I encountered the "first-run slow-down" problem with GTX 1080 cards and nvidia-docker as discussed in this question. I'm using the TensorFlow build from its official pip package and a custom docker image based on nvidia-docker's Ubuntu 16.04 base image. How do I make TensorFlow to load (and build JIT caches) all registered CUDA kernels programmatically in a Dockerfile? (rather than manually building TensorFlow using TF_CUDA_COMPUTE_CAPABILITIES environment variable) 回答1: There seems to be no