google-research/deeplab2

How to identify model is using GPU ?

PritiDrishtic opened this issue · 0 comments

I exported the KMAX model using export_model.py on GPU( Tesla T4) .

Please advise how I can determine whether this model is using GPU as the execution performance for pretrained model from CPU mode and this model is nearly identical.

executing inference on an image and attaching tensorflow logs from script -

I tensorflow/core/common_runtime/executor.cc:1197] [/device:CPU:0] (DEBUG INFO) Executor start aborting (this does not indicate an error and you can ignore this message): INVALID_ARGUMENT: You must feed a value for placeholder tensor 'input_tensor' with dtype uint8 and shape [1530,2720,3]      [[{{node input_tensor}}]] 2023-05-17 05:35:42.665202: I tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:424] Loaded cuDNN version 8800 2023-05-17 05:35:43.606321: I tensorflow/compiler/xla/service/service.cc:169] XLA service 0x7f29702d1dc0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices: 2023-05-17 05:35:43.606368: I tensorflow/compiler/xla/service/service.cc:177]   StreamExecutor device (0): Tesla T4, Compute Capability 7.5 2023-05-17 05:35:44.473843: I ./tensorflow/compiler/jit/device_compiler.h:180] Compiled cluster using XLA!  This line is logged at most once for the lifetime of the process.