Package ai.djl.serving.wlm
package ai.djl.serving.wlm
Contains the model server backend which manages worker threads and executes jobs on models.
- See Also:
-
ClassDescriptionAn adapter is a modification producing a variation of a model that can be used during prediction.Job<I,
O> A class represents an inference job.JobFunction<I,O> A function describing the action to take in aJob
.A utility class to auto configure LMI model properties.A utility class to detect optimal engine for LMI model.ModelInfo<I,O> A class represent a loaded model and it's metadata.a batch aggregator that never terminates by itself.An overload ofAdapter
for the python engine.A utility class to detect optimal engine for SageMaker saved model.a batch aggregator that terminates after a maximum idle time.WorkerGroup<I,O> WorkerPool<I,O> Manages the work load for a single model.WorkerPoolConfig<I,O> AWorkerPoolConfig
represents a task that could be run in theWorkLoadManager
.An enum represents state of a worker type.The part of theWorkerPoolConfig
for an individualWorkerThread
.An enum represents state of a worker.WorkerThread<I,O> TheWorkerThread
is the worker managed by theWorkLoadManager
.WorkerThread.Builder<I,O> A Builder to construct aWorkerThread
.WorkLoadManager is responsible to manage the work load of worker thread.