class PredictionService[T] extends AnyRef

Thread-safe Prediction Service for Concurrent Calls

In this service, concurrency is kept not greater than numThreads by a BlockingQueue, which contains available model instances.

numThreads model instances sharing weights/bias will be put into the BlockingQueue during initialization.

When predict method called, service will try to take an instance from BlockingQueue, which means if all instances are on serving, the predicting request will be blocked until some instances are released.

If exceptions caught during predict, a scalar Tensor[String] will be returned with thrown message.

Linear Supertypes
AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. PredictionService
  2. AnyRef
  3. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def clone(): AnyRef
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @native() @throws( ... )
  6. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  7. def equals(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  8. def finalize(): Unit
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  9. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  10. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  11. val instQueue: LinkedBlockingQueue[Module[T]]
    Attributes
    protected
  12. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  13. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  14. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  15. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  16. def predict(request: Array[Byte]): Array[Byte]

    Firstly, deserialization tasks will be run with inputs(Array[Byte]).

    Thread-safe single sample prediction

    Firstly, deserialization tasks will be run with inputs(Array[Byte]).

    Then, run model prediction with deserialized inputs as soon as there exists vacant instances(total number is numThreads). Otherwise, it will hold on till some instances are released.

    Finally, prediction results will be serialized to Array[Byte] according to BigDL.proto.

    request

    input bytes, which will be deserialized by BigDL.proto

    returns

    output bytes, which is serialized by BigDl.proto

  17. def predict(request: Activity): Activity

    Running model prediction with input Activity as soon as there exists vacant instances(the size of pool is numThreads).

    Thread-safe single sample prediction

    Running model prediction with input Activity as soon as there exists vacant instances(the size of pool is numThreads). Otherwise, it will hold on till some instances are released.

    Outputs will be deeply copied after model prediction, so they are invariant.

    request

    input Activity, could be Tensor or Table(key, Tensor)

    returns

    output Activity, could be Tensor or Table(key, Tensor)

  18. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  19. def toString(): String
    Definition Classes
    AnyRef → Any
  20. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  21. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  22. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @throws( ... )

Inherited from AnyRef

Inherited from Any

Ungrouped