TensorFlow Serving是一个针对机器学习模型的灵活,高性能的服务系统,专为生产环境而设计。使用TensorFlow Serving可以轻松部署新算法和实验,同时保持相同的服务器体系结构和API。
这里不对此过多介绍,具体可以查看官方文档:https://www.tensorflow.org/tfx/guide/serving
工作的时候部署一个人脸识别服务上线,模型训练好之后使用TensorFlow Serving部署
Flags:
--port=8500 int32 Port to listen on for gRPC API
--grpc_socket_path="" string If non-empty, listen to a UNIX socket for gRPC API on the given path. Can be either relative or absolute path.
--rest_api_port=0 int32 Port to listen on for HTTP/REST API. If set to zero HTTP/REST API will not be exported. This port must be different than the one specified in --port.
--rest_api_num_threads=16 int32 Number of threads for HTTP/REST API processing. If not set, will be auto set based on number of CPUs.
--rest_api_timeout_in_ms=30000 int32 Timeout for HTTP/REST API calls.
--enable_batching=false bool enable batching
--batching_parameters_file="" string If non-empty, read an ascii BatchingParameters protobuf from the supplied file name and use the contained values instead of the defaults.
--model_config_file="" string If non-