ts package¶
Subpackages¶
- ts.metrics package
- ts.model_service package
- ts.protocol package
- ts.torch_handler package
- Subpackages
- ts.torch_handler.request_envelope package
- ts.torch_handler.unit_tests package
- Subpackages
- Submodules
- ts.torch_handler.unit_tests.test_base_handler module
- ts.torch_handler.unit_tests.test_envelopes module
- ts.torch_handler.unit_tests.test_image_classifier module
- ts.torch_handler.unit_tests.test_image_segmenter module
- ts.torch_handler.unit_tests.test_mnist_kf module
- ts.torch_handler.unit_tests.test_object_detector module
- Module contents
- Submodules
- ts.torch_handler.base_handler module
- ts.torch_handler.contractions module
- ts.torch_handler.densenet_handler module
- ts.torch_handler.image_classifier module
- ts.torch_handler.image_segmenter module
- ts.torch_handler.object_detector module
- ts.torch_handler.text_classifier module
- ts.torch_handler.text_handler module
- ts.torch_handler.vision_handler module
- Module contents
- Subpackages
- ts.utils package
Submodules¶
ts.arg_parser module¶
This module parses the arguments given through the torchserve command-line. This is used by model-server at runtime.
ts.context module¶
Context object of incoming request
- class ts.context.Context(model_name, model_dir, manifest, batch_size, gpu, mms_version, limit_max_image_pixels=True, metrics=None, model_yaml_config=None)[source]¶
Bases:
object
Context stores model relevant worker information Some fixed during load times and some set by the service
- property metrics¶
- property request_processor¶
- set_all_response_status(code: int = 200, phrase: str = '') None [source]¶
Set the status code of individual requests :param phrase: :param code: :return:
- set_response_status(code: int = 200, phrase: str = '', idx: int = 0)[source]¶
Set the status code of individual requests :param phrase: :param idx: The index data in the list(data) that is sent to the handle() method :param code: :return:
- property system_properties¶
ts.model_loader module¶
Model loader.
- class ts.model_loader.ModelLoader[source]¶
Bases:
object
Base Model Loader class.
- abstract load(model_name: str, model_dir: str, handler: Optional[str] = None, gpu_id: Optional[int] = None, batch_size: Optional[int] = None, envelope: Optional[str] = None, limit_max_image_pixels: Optional[bool] = True)[source]¶
Load model from file.
- Parameters:
model_name –
model_dir –
handler –
gpu_id –
batch_size –
envelope –
limit_max_image_pixels –
- Returns:
Model
- class ts.model_loader.TsModelLoader[source]¶
Bases:
ModelLoader
TorchServe 1.0 Model Loader
- load(model_name: str, model_dir: str, handler: Optional[str] = None, gpu_id: Optional[int] = None, batch_size: Optional[int] = None, envelope: Optional[str] = None, limit_max_image_pixels: Optional[bool] = True, metrics_cache: Optional[MetricsCacheYamlImpl] = None) Service [source]¶
Load TorchServe 1.0 model from file.
- Parameters:
model_name –
model_dir –
handler –
gpu_id –
batch_size –
envelope –
limit_max_image_pixels –
metrics_cache – MetricsCacheYamlImpl object
- Returns:
ts.model_server module¶
File to define the entry point to Model Server
ts.model_service_worker module¶
ModelServiceWorker is the worker that is started by the MMS front-end. Communication message format: binary encoding
- class ts.model_service_worker.TorchModelServiceWorker(s_type: Optional[str] = None, s_name: Optional[str] = None, host_addr: Optional[str] = None, port_num: Optional[int] = None, metrics_config: Optional[str] = None, async_comm: Optional[bool] = False)[source]¶
Bases:
object
Backend worker to handle Model Server’s python service code
- handle_connection_async(cl_socket)[source]¶
Handle socket connection.
- Parameters:
cl_socket –
- Returns:
- load_model(load_model_request)[source]¶
Expected command {
“command” : “load”, string “modelPath” : “/path/to/model/file”, string “modelName” : “name”, string “gpu” : None if CPU else gpu_id, int “handler” : service handler entry point if provided, string “envelope” : name of wrapper/unwrapper of request data if provided, string “batchSize” : batch size, int “limitMaxImagePixels”: limit pillow image max_image_pixels, bool
}
- Parameters:
load_model_request –
- Returns:
ts.service module¶
CustomService class definitions
- class ts.service.Service(model_name, model_dir, manifest, entry_point, gpu, batch_size, limit_max_image_pixels=True, metrics_cache=None)[source]¶
Bases:
object
Wrapper for custom entry_point
- property context¶
- predict(batch)[source]¶
- PREDICT COMMAND = {
“command”: “predict”, “batch”: [ REQUEST_INPUT ]
} :param batch: list of request :return:
ts.version module¶
This is the current version of TorchServe
Module contents¶
This module does the following: a. Starts model-server. b. Creates end-points based on the configured models. c. Exposes standard “ping” and “api-description” endpoints. d. Waits for servicing inference requests.