tempo.mlserver module¶ class tempo.mlserver.InferenceRuntime(settings: mlserver.settings.ModelSettings)¶ Bases: mlserver.model.MLModel async load() → bool¶ async predict(request: mlserver.types.dataplane.InferenceRequest) → mlserver.types.dataplane.InferenceResponse¶