MIVisionX Inference Server
This Sample Inference Server supports:
- convert and maintain a database of pre-trained CAFFE models using Model Compiler
- allow multiple TCP/IP client connections for inference work submissions
- multi-GPU high-throughput live streaming batch scheduler
Command-line usage:
inference_server_app [-p <port> default:26262]
[-b <batch size> default:64]
[-n <model compiler path> default:/opt/rocm/libexec/mivisionx/model_compiler/python]
[-fp16 <ON:1 or OFF:0> default:0]
[-w <server working directory> default:~/]
[-t <num cpu decoder threads [2-64]> default:1]
[-gpu <comma separated list of GPUs>]
[-q <max pending batches>]
[-s <local shadow folder full path>]
Make sure that all executables and libraries are in PATH
and LD_LIBRARY_PATH
environment variables.
% export PATH=$PATH:/opt/rocm/bin
% export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/opt/rocm/lib
The inference_server_app
works with Client Application.
- Execute
inference_server_app
on the server machine with Radeon Instinct GPUs - Execute
Client Application
on one or more workstations: connect to the server and classify images using any pre-trained neural network