public static interface PredictionServiceGrpc.AsyncService
A service for online predictions and explanations.
| Modifier and Type | Method and Description |
|---|---|
default void |
countTokens(CountTokensRequest request,
io.grpc.stub.StreamObserver<CountTokensResponse> responseObserver)
Perform a token counting.
|
default void |
directPredict(DirectPredictRequest request,
io.grpc.stub.StreamObserver<DirectPredictResponse> responseObserver)
Perform an unary online prediction request for Vertex first-party products
and frameworks.
|
default void |
directRawPredict(DirectRawPredictRequest request,
io.grpc.stub.StreamObserver<DirectRawPredictResponse> responseObserver)
Perform an online prediction request through gRPC.
|
default void |
explain(ExplainRequest request,
io.grpc.stub.StreamObserver<ExplainResponse> responseObserver)
Perform an online explanation.
|
default void |
predict(PredictRequest request,
io.grpc.stub.StreamObserver<PredictResponse> responseObserver)
Perform an online prediction.
|
default void |
rawPredict(RawPredictRequest request,
io.grpc.stub.StreamObserver<com.google.api.HttpBody> responseObserver)
Perform an online prediction with an arbitrary HTTP payload.
|
default void |
serverStreamingPredict(StreamingPredictRequest request,
io.grpc.stub.StreamObserver<StreamingPredictResponse> responseObserver)
Perform a server-side streaming online prediction request for Vertex
LLM streaming.
|
default void |
streamGenerateContent(GenerateContentRequest request,
io.grpc.stub.StreamObserver<GenerateContentResponse> responseObserver)
Generate content with multimodal inputs with streaming support.
|
default io.grpc.stub.StreamObserver<StreamingPredictRequest> |
streamingPredict(io.grpc.stub.StreamObserver<StreamingPredictResponse> responseObserver)
Perform a streaming online prediction request for Vertex first-party
products and frameworks.
|
default io.grpc.stub.StreamObserver<StreamingRawPredictRequest> |
streamingRawPredict(io.grpc.stub.StreamObserver<StreamingRawPredictResponse> responseObserver)
Perform a streaming online prediction request through gRPC.
|
default void predict(PredictRequest request, io.grpc.stub.StreamObserver<PredictResponse> responseObserver)
Perform an online prediction.
default void rawPredict(RawPredictRequest request, io.grpc.stub.StreamObserver<com.google.api.HttpBody> responseObserver)
Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the [Endpoint][google.cloud.aiplatform.v1beta1.Endpoint] that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's [DeployedModel][google.cloud.aiplatform.v1beta1.DeployedModel] that served this prediction.
default void directPredict(DirectPredictRequest request, io.grpc.stub.StreamObserver<DirectPredictResponse> responseObserver)
Perform an unary online prediction request for Vertex first-party products and frameworks.
default void directRawPredict(DirectRawPredictRequest request, io.grpc.stub.StreamObserver<DirectRawPredictResponse> responseObserver)
Perform an online prediction request through gRPC.
default io.grpc.stub.StreamObserver<StreamingPredictRequest> streamingPredict(io.grpc.stub.StreamObserver<StreamingPredictResponse> responseObserver)
Perform a streaming online prediction request for Vertex first-party products and frameworks.
default void serverStreamingPredict(StreamingPredictRequest request, io.grpc.stub.StreamObserver<StreamingPredictResponse> responseObserver)
Perform a server-side streaming online prediction request for Vertex LLM streaming.
default io.grpc.stub.StreamObserver<StreamingRawPredictRequest> streamingRawPredict(io.grpc.stub.StreamObserver<StreamingRawPredictResponse> responseObserver)
Perform a streaming online prediction request through gRPC.
default void explain(ExplainRequest request, io.grpc.stub.StreamObserver<ExplainResponse> responseObserver)
Perform an online explanation. If [deployed_model_id][google.cloud.aiplatform.v1beta1.ExplainRequest.deployed_model_id] is specified, the corresponding DeployModel must have [explanation_spec][google.cloud.aiplatform.v1beta1.DeployedModel.explanation_spec] populated. If [deployed_model_id][google.cloud.aiplatform.v1beta1.ExplainRequest.deployed_model_id] is not specified, all DeployedModels must have [explanation_spec][google.cloud.aiplatform.v1beta1.DeployedModel.explanation_spec] populated.
default void countTokens(CountTokensRequest request, io.grpc.stub.StreamObserver<CountTokensResponse> responseObserver)
Perform a token counting.
default void streamGenerateContent(GenerateContentRequest request, io.grpc.stub.StreamObserver<GenerateContentResponse> responseObserver)
Generate content with multimodal inputs with streaming support.
Copyright © 2024 Google LLC. All rights reserved.