PatWie/tensorflow-cmake

how to set gpu configuration

AllenDun opened this issue · 1 comments

Environment

If an issue occurs it is helpful to provide several information: Please replace these information below:

info output
os uname -a
c++ version c++ --version
TF version python -c "import tensorflow as tf; print(tf.__version__);"
TF is working python -c "import tensorflow as tf; sess=tf.InteractiveSession()" works

You might run the tf_info.sh script in your terminal to gather these information and manually anonymize the output when necessary before pasting it here.

Issue

A clear and concise description of what the bug is will be helpful to resolve issues.

Context:
Please explain the conditions which led you to write this issue, eg. "writing a custom op" or "doing inference on device given a keras model".

Reproduce:
Please explain what you did (exactly), eg., files you modified, or, commands you entered so that I can reproduce if possible.

Output
Please explain what you get as an result or output from the reproduce-steps above, eg. "I get the output/error message ... during the compilation". Please provide the exact output and no interpretation.

Expectation
Please explain what you expected to get, eg. "the inference code x should produce y".

Investigation
If possible, please explain, which steps you did to investigate this issues. What are your findings.

nice work! could you add some codes in your examples to support gpu configurations for multi-gpu environment? It is really bothering when do inference with tensorflow in multi-gpu machine.Thanks

Have you tried

CUDA_VISIBLE_DEVICES=0 ./path/to/script