All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.google.cloud.automl.v1.PredictionServiceGrpc Maven / Gradle / Ivy

There is a newer version: 2.54.0
Show newest version
/*
 * Copyright 2024 Google LLC
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *     https://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package com.google.cloud.automl.v1;

import static io.grpc.MethodDescriptor.generateFullMethodName;

/**
 *
 *
 * 
 * AutoML Prediction API.
 * On any input that is documented to expect a string parameter in
 * snake_case or dash-case, either of those cases is accepted.
 * 
*/ @javax.annotation.Generated( value = "by gRPC proto compiler", comments = "Source: google/cloud/automl/v1/prediction_service.proto") @io.grpc.stub.annotations.GrpcGenerated public final class PredictionServiceGrpc { private PredictionServiceGrpc() {} public static final java.lang.String SERVICE_NAME = "google.cloud.automl.v1.PredictionService"; // Static method descriptors that strictly reflect the proto. private static volatile io.grpc.MethodDescriptor< com.google.cloud.automl.v1.PredictRequest, com.google.cloud.automl.v1.PredictResponse> getPredictMethod; @io.grpc.stub.annotations.RpcMethod( fullMethodName = SERVICE_NAME + '/' + "Predict", requestType = com.google.cloud.automl.v1.PredictRequest.class, responseType = com.google.cloud.automl.v1.PredictResponse.class, methodType = io.grpc.MethodDescriptor.MethodType.UNARY) public static io.grpc.MethodDescriptor< com.google.cloud.automl.v1.PredictRequest, com.google.cloud.automl.v1.PredictResponse> getPredictMethod() { io.grpc.MethodDescriptor< com.google.cloud.automl.v1.PredictRequest, com.google.cloud.automl.v1.PredictResponse> getPredictMethod; if ((getPredictMethod = PredictionServiceGrpc.getPredictMethod) == null) { synchronized (PredictionServiceGrpc.class) { if ((getPredictMethod = PredictionServiceGrpc.getPredictMethod) == null) { PredictionServiceGrpc.getPredictMethod = getPredictMethod = io.grpc.MethodDescriptor . newBuilder() .setType(io.grpc.MethodDescriptor.MethodType.UNARY) .setFullMethodName(generateFullMethodName(SERVICE_NAME, "Predict")) .setSampledToLocalTracing(true) .setRequestMarshaller( io.grpc.protobuf.ProtoUtils.marshaller( com.google.cloud.automl.v1.PredictRequest.getDefaultInstance())) .setResponseMarshaller( io.grpc.protobuf.ProtoUtils.marshaller( com.google.cloud.automl.v1.PredictResponse.getDefaultInstance())) .setSchemaDescriptor(new PredictionServiceMethodDescriptorSupplier("Predict")) .build(); } } } return getPredictMethod; } private static volatile io.grpc.MethodDescriptor< com.google.cloud.automl.v1.BatchPredictRequest, com.google.longrunning.Operation> getBatchPredictMethod; @io.grpc.stub.annotations.RpcMethod( fullMethodName = SERVICE_NAME + '/' + "BatchPredict", requestType = com.google.cloud.automl.v1.BatchPredictRequest.class, responseType = com.google.longrunning.Operation.class, methodType = io.grpc.MethodDescriptor.MethodType.UNARY) public static io.grpc.MethodDescriptor< com.google.cloud.automl.v1.BatchPredictRequest, com.google.longrunning.Operation> getBatchPredictMethod() { io.grpc.MethodDescriptor< com.google.cloud.automl.v1.BatchPredictRequest, com.google.longrunning.Operation> getBatchPredictMethod; if ((getBatchPredictMethod = PredictionServiceGrpc.getBatchPredictMethod) == null) { synchronized (PredictionServiceGrpc.class) { if ((getBatchPredictMethod = PredictionServiceGrpc.getBatchPredictMethod) == null) { PredictionServiceGrpc.getBatchPredictMethod = getBatchPredictMethod = io.grpc.MethodDescriptor . newBuilder() .setType(io.grpc.MethodDescriptor.MethodType.UNARY) .setFullMethodName(generateFullMethodName(SERVICE_NAME, "BatchPredict")) .setSampledToLocalTracing(true) .setRequestMarshaller( io.grpc.protobuf.ProtoUtils.marshaller( com.google.cloud.automl.v1.BatchPredictRequest.getDefaultInstance())) .setResponseMarshaller( io.grpc.protobuf.ProtoUtils.marshaller( com.google.longrunning.Operation.getDefaultInstance())) .setSchemaDescriptor( new PredictionServiceMethodDescriptorSupplier("BatchPredict")) .build(); } } } return getBatchPredictMethod; } /** Creates a new async stub that supports all call types for the service */ public static PredictionServiceStub newStub(io.grpc.Channel channel) { io.grpc.stub.AbstractStub.StubFactory factory = new io.grpc.stub.AbstractStub.StubFactory() { @java.lang.Override public PredictionServiceStub newStub( io.grpc.Channel channel, io.grpc.CallOptions callOptions) { return new PredictionServiceStub(channel, callOptions); } }; return PredictionServiceStub.newStub(factory, channel); } /** * Creates a new blocking-style stub that supports unary and streaming output calls on the service */ public static PredictionServiceBlockingStub newBlockingStub(io.grpc.Channel channel) { io.grpc.stub.AbstractStub.StubFactory factory = new io.grpc.stub.AbstractStub.StubFactory() { @java.lang.Override public PredictionServiceBlockingStub newStub( io.grpc.Channel channel, io.grpc.CallOptions callOptions) { return new PredictionServiceBlockingStub(channel, callOptions); } }; return PredictionServiceBlockingStub.newStub(factory, channel); } /** Creates a new ListenableFuture-style stub that supports unary calls on the service */ public static PredictionServiceFutureStub newFutureStub(io.grpc.Channel channel) { io.grpc.stub.AbstractStub.StubFactory factory = new io.grpc.stub.AbstractStub.StubFactory() { @java.lang.Override public PredictionServiceFutureStub newStub( io.grpc.Channel channel, io.grpc.CallOptions callOptions) { return new PredictionServiceFutureStub(channel, callOptions); } }; return PredictionServiceFutureStub.newStub(factory, channel); } /** * * *
   * AutoML Prediction API.
   * On any input that is documented to expect a string parameter in
   * snake_case or dash-case, either of those cases is accepted.
   * 
*/ public interface AsyncService { /** * * *
     * Perform an online prediction. The prediction result is directly
     * returned in the response.
     * Available for following ML scenarios, and their expected request payloads:
     * AutoML Vision Classification
     * * An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
     * AutoML Vision Object Detection
     * * An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
     * AutoML Natural Language Classification
     * * A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
     * .PDF, .TIF or .TIFF format with size upto 2MB.
     * AutoML Natural Language Entity Extraction
     * * A TextSnippet up to 10,000 characters, UTF-8 NFC encoded or a document
     *  in .PDF, .TIF or .TIFF format with size upto 20MB.
     * AutoML Natural Language Sentiment Analysis
     * * A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
     * .PDF, .TIF or .TIFF format with size upto 2MB.
     * AutoML Translation
     * * A TextSnippet up to 25,000 characters, UTF-8 encoded.
     * AutoML Tables
     * * A row with column values matching
     *   the columns of the model, up to 5MB. Not available for FORECASTING
     *   `prediction_type`.
     * 
*/ default void predict( com.google.cloud.automl.v1.PredictRequest request, io.grpc.stub.StreamObserver responseObserver) { io.grpc.stub.ServerCalls.asyncUnimplementedUnaryCall(getPredictMethod(), responseObserver); } /** * * *
     * Perform a batch prediction. Unlike the online [Predict][google.cloud.automl.v1.PredictionService.Predict], batch
     * prediction result won't be immediately available in the response. Instead,
     * a long running operation object is returned. User can poll the operation
     * result via [GetOperation][google.longrunning.Operations.GetOperation]
     * method. Once the operation is done, [BatchPredictResult][google.cloud.automl.v1.BatchPredictResult] is returned in
     * the [response][google.longrunning.Operation.response] field.
     * Available for following ML scenarios:
     * * AutoML Vision Classification
     * * AutoML Vision Object Detection
     * * AutoML Video Intelligence Classification
     * * AutoML Video Intelligence Object Tracking * AutoML Natural Language Classification
     * * AutoML Natural Language Entity Extraction
     * * AutoML Natural Language Sentiment Analysis
     * * AutoML Tables
     * 
*/ default void batchPredict( com.google.cloud.automl.v1.BatchPredictRequest request, io.grpc.stub.StreamObserver responseObserver) { io.grpc.stub.ServerCalls.asyncUnimplementedUnaryCall( getBatchPredictMethod(), responseObserver); } } /** * Base class for the server implementation of the service PredictionService. * *
   * AutoML Prediction API.
   * On any input that is documented to expect a string parameter in
   * snake_case or dash-case, either of those cases is accepted.
   * 
*/ public abstract static class PredictionServiceImplBase implements io.grpc.BindableService, AsyncService { @java.lang.Override public final io.grpc.ServerServiceDefinition bindService() { return PredictionServiceGrpc.bindService(this); } } /** * A stub to allow clients to do asynchronous rpc calls to service PredictionService. * *
   * AutoML Prediction API.
   * On any input that is documented to expect a string parameter in
   * snake_case or dash-case, either of those cases is accepted.
   * 
*/ public static final class PredictionServiceStub extends io.grpc.stub.AbstractAsyncStub { private PredictionServiceStub(io.grpc.Channel channel, io.grpc.CallOptions callOptions) { super(channel, callOptions); } @java.lang.Override protected PredictionServiceStub build( io.grpc.Channel channel, io.grpc.CallOptions callOptions) { return new PredictionServiceStub(channel, callOptions); } /** * * *
     * Perform an online prediction. The prediction result is directly
     * returned in the response.
     * Available for following ML scenarios, and their expected request payloads:
     * AutoML Vision Classification
     * * An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
     * AutoML Vision Object Detection
     * * An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
     * AutoML Natural Language Classification
     * * A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
     * .PDF, .TIF or .TIFF format with size upto 2MB.
     * AutoML Natural Language Entity Extraction
     * * A TextSnippet up to 10,000 characters, UTF-8 NFC encoded or a document
     *  in .PDF, .TIF or .TIFF format with size upto 20MB.
     * AutoML Natural Language Sentiment Analysis
     * * A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
     * .PDF, .TIF or .TIFF format with size upto 2MB.
     * AutoML Translation
     * * A TextSnippet up to 25,000 characters, UTF-8 encoded.
     * AutoML Tables
     * * A row with column values matching
     *   the columns of the model, up to 5MB. Not available for FORECASTING
     *   `prediction_type`.
     * 
*/ public void predict( com.google.cloud.automl.v1.PredictRequest request, io.grpc.stub.StreamObserver responseObserver) { io.grpc.stub.ClientCalls.asyncUnaryCall( getChannel().newCall(getPredictMethod(), getCallOptions()), request, responseObserver); } /** * * *
     * Perform a batch prediction. Unlike the online [Predict][google.cloud.automl.v1.PredictionService.Predict], batch
     * prediction result won't be immediately available in the response. Instead,
     * a long running operation object is returned. User can poll the operation
     * result via [GetOperation][google.longrunning.Operations.GetOperation]
     * method. Once the operation is done, [BatchPredictResult][google.cloud.automl.v1.BatchPredictResult] is returned in
     * the [response][google.longrunning.Operation.response] field.
     * Available for following ML scenarios:
     * * AutoML Vision Classification
     * * AutoML Vision Object Detection
     * * AutoML Video Intelligence Classification
     * * AutoML Video Intelligence Object Tracking * AutoML Natural Language Classification
     * * AutoML Natural Language Entity Extraction
     * * AutoML Natural Language Sentiment Analysis
     * * AutoML Tables
     * 
*/ public void batchPredict( com.google.cloud.automl.v1.BatchPredictRequest request, io.grpc.stub.StreamObserver responseObserver) { io.grpc.stub.ClientCalls.asyncUnaryCall( getChannel().newCall(getBatchPredictMethod(), getCallOptions()), request, responseObserver); } } /** * A stub to allow clients to do synchronous rpc calls to service PredictionService. * *
   * AutoML Prediction API.
   * On any input that is documented to expect a string parameter in
   * snake_case or dash-case, either of those cases is accepted.
   * 
*/ public static final class PredictionServiceBlockingStub extends io.grpc.stub.AbstractBlockingStub { private PredictionServiceBlockingStub( io.grpc.Channel channel, io.grpc.CallOptions callOptions) { super(channel, callOptions); } @java.lang.Override protected PredictionServiceBlockingStub build( io.grpc.Channel channel, io.grpc.CallOptions callOptions) { return new PredictionServiceBlockingStub(channel, callOptions); } /** * * *
     * Perform an online prediction. The prediction result is directly
     * returned in the response.
     * Available for following ML scenarios, and their expected request payloads:
     * AutoML Vision Classification
     * * An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
     * AutoML Vision Object Detection
     * * An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
     * AutoML Natural Language Classification
     * * A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
     * .PDF, .TIF or .TIFF format with size upto 2MB.
     * AutoML Natural Language Entity Extraction
     * * A TextSnippet up to 10,000 characters, UTF-8 NFC encoded or a document
     *  in .PDF, .TIF or .TIFF format with size upto 20MB.
     * AutoML Natural Language Sentiment Analysis
     * * A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
     * .PDF, .TIF or .TIFF format with size upto 2MB.
     * AutoML Translation
     * * A TextSnippet up to 25,000 characters, UTF-8 encoded.
     * AutoML Tables
     * * A row with column values matching
     *   the columns of the model, up to 5MB. Not available for FORECASTING
     *   `prediction_type`.
     * 
*/ public com.google.cloud.automl.v1.PredictResponse predict( com.google.cloud.automl.v1.PredictRequest request) { return io.grpc.stub.ClientCalls.blockingUnaryCall( getChannel(), getPredictMethod(), getCallOptions(), request); } /** * * *
     * Perform a batch prediction. Unlike the online [Predict][google.cloud.automl.v1.PredictionService.Predict], batch
     * prediction result won't be immediately available in the response. Instead,
     * a long running operation object is returned. User can poll the operation
     * result via [GetOperation][google.longrunning.Operations.GetOperation]
     * method. Once the operation is done, [BatchPredictResult][google.cloud.automl.v1.BatchPredictResult] is returned in
     * the [response][google.longrunning.Operation.response] field.
     * Available for following ML scenarios:
     * * AutoML Vision Classification
     * * AutoML Vision Object Detection
     * * AutoML Video Intelligence Classification
     * * AutoML Video Intelligence Object Tracking * AutoML Natural Language Classification
     * * AutoML Natural Language Entity Extraction
     * * AutoML Natural Language Sentiment Analysis
     * * AutoML Tables
     * 
*/ public com.google.longrunning.Operation batchPredict( com.google.cloud.automl.v1.BatchPredictRequest request) { return io.grpc.stub.ClientCalls.blockingUnaryCall( getChannel(), getBatchPredictMethod(), getCallOptions(), request); } } /** * A stub to allow clients to do ListenableFuture-style rpc calls to service PredictionService. * *
   * AutoML Prediction API.
   * On any input that is documented to expect a string parameter in
   * snake_case or dash-case, either of those cases is accepted.
   * 
*/ public static final class PredictionServiceFutureStub extends io.grpc.stub.AbstractFutureStub { private PredictionServiceFutureStub(io.grpc.Channel channel, io.grpc.CallOptions callOptions) { super(channel, callOptions); } @java.lang.Override protected PredictionServiceFutureStub build( io.grpc.Channel channel, io.grpc.CallOptions callOptions) { return new PredictionServiceFutureStub(channel, callOptions); } /** * * *
     * Perform an online prediction. The prediction result is directly
     * returned in the response.
     * Available for following ML scenarios, and their expected request payloads:
     * AutoML Vision Classification
     * * An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
     * AutoML Vision Object Detection
     * * An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
     * AutoML Natural Language Classification
     * * A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
     * .PDF, .TIF or .TIFF format with size upto 2MB.
     * AutoML Natural Language Entity Extraction
     * * A TextSnippet up to 10,000 characters, UTF-8 NFC encoded or a document
     *  in .PDF, .TIF or .TIFF format with size upto 20MB.
     * AutoML Natural Language Sentiment Analysis
     * * A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
     * .PDF, .TIF or .TIFF format with size upto 2MB.
     * AutoML Translation
     * * A TextSnippet up to 25,000 characters, UTF-8 encoded.
     * AutoML Tables
     * * A row with column values matching
     *   the columns of the model, up to 5MB. Not available for FORECASTING
     *   `prediction_type`.
     * 
*/ public com.google.common.util.concurrent.ListenableFuture< com.google.cloud.automl.v1.PredictResponse> predict(com.google.cloud.automl.v1.PredictRequest request) { return io.grpc.stub.ClientCalls.futureUnaryCall( getChannel().newCall(getPredictMethod(), getCallOptions()), request); } /** * * *
     * Perform a batch prediction. Unlike the online [Predict][google.cloud.automl.v1.PredictionService.Predict], batch
     * prediction result won't be immediately available in the response. Instead,
     * a long running operation object is returned. User can poll the operation
     * result via [GetOperation][google.longrunning.Operations.GetOperation]
     * method. Once the operation is done, [BatchPredictResult][google.cloud.automl.v1.BatchPredictResult] is returned in
     * the [response][google.longrunning.Operation.response] field.
     * Available for following ML scenarios:
     * * AutoML Vision Classification
     * * AutoML Vision Object Detection
     * * AutoML Video Intelligence Classification
     * * AutoML Video Intelligence Object Tracking * AutoML Natural Language Classification
     * * AutoML Natural Language Entity Extraction
     * * AutoML Natural Language Sentiment Analysis
     * * AutoML Tables
     * 
*/ public com.google.common.util.concurrent.ListenableFuture batchPredict(com.google.cloud.automl.v1.BatchPredictRequest request) { return io.grpc.stub.ClientCalls.futureUnaryCall( getChannel().newCall(getBatchPredictMethod(), getCallOptions()), request); } } private static final int METHODID_PREDICT = 0; private static final int METHODID_BATCH_PREDICT = 1; private static final class MethodHandlers implements io.grpc.stub.ServerCalls.UnaryMethod, io.grpc.stub.ServerCalls.ServerStreamingMethod, io.grpc.stub.ServerCalls.ClientStreamingMethod, io.grpc.stub.ServerCalls.BidiStreamingMethod { private final AsyncService serviceImpl; private final int methodId; MethodHandlers(AsyncService serviceImpl, int methodId) { this.serviceImpl = serviceImpl; this.methodId = methodId; } @java.lang.Override @java.lang.SuppressWarnings("unchecked") public void invoke(Req request, io.grpc.stub.StreamObserver responseObserver) { switch (methodId) { case METHODID_PREDICT: serviceImpl.predict( (com.google.cloud.automl.v1.PredictRequest) request, (io.grpc.stub.StreamObserver) responseObserver); break; case METHODID_BATCH_PREDICT: serviceImpl.batchPredict( (com.google.cloud.automl.v1.BatchPredictRequest) request, (io.grpc.stub.StreamObserver) responseObserver); break; default: throw new AssertionError(); } } @java.lang.Override @java.lang.SuppressWarnings("unchecked") public io.grpc.stub.StreamObserver invoke( io.grpc.stub.StreamObserver responseObserver) { switch (methodId) { default: throw new AssertionError(); } } } public static final io.grpc.ServerServiceDefinition bindService(AsyncService service) { return io.grpc.ServerServiceDefinition.builder(getServiceDescriptor()) .addMethod( getPredictMethod(), io.grpc.stub.ServerCalls.asyncUnaryCall( new MethodHandlers< com.google.cloud.automl.v1.PredictRequest, com.google.cloud.automl.v1.PredictResponse>(service, METHODID_PREDICT))) .addMethod( getBatchPredictMethod(), io.grpc.stub.ServerCalls.asyncUnaryCall( new MethodHandlers< com.google.cloud.automl.v1.BatchPredictRequest, com.google.longrunning.Operation>(service, METHODID_BATCH_PREDICT))) .build(); } private abstract static class PredictionServiceBaseDescriptorSupplier implements io.grpc.protobuf.ProtoFileDescriptorSupplier, io.grpc.protobuf.ProtoServiceDescriptorSupplier { PredictionServiceBaseDescriptorSupplier() {} @java.lang.Override public com.google.protobuf.Descriptors.FileDescriptor getFileDescriptor() { return com.google.cloud.automl.v1.PredictionServiceProto.getDescriptor(); } @java.lang.Override public com.google.protobuf.Descriptors.ServiceDescriptor getServiceDescriptor() { return getFileDescriptor().findServiceByName("PredictionService"); } } private static final class PredictionServiceFileDescriptorSupplier extends PredictionServiceBaseDescriptorSupplier { PredictionServiceFileDescriptorSupplier() {} } private static final class PredictionServiceMethodDescriptorSupplier extends PredictionServiceBaseDescriptorSupplier implements io.grpc.protobuf.ProtoMethodDescriptorSupplier { private final java.lang.String methodName; PredictionServiceMethodDescriptorSupplier(java.lang.String methodName) { this.methodName = methodName; } @java.lang.Override public com.google.protobuf.Descriptors.MethodDescriptor getMethodDescriptor() { return getServiceDescriptor().findMethodByName(methodName); } } private static volatile io.grpc.ServiceDescriptor serviceDescriptor; public static io.grpc.ServiceDescriptor getServiceDescriptor() { io.grpc.ServiceDescriptor result = serviceDescriptor; if (result == null) { synchronized (PredictionServiceGrpc.class) { result = serviceDescriptor; if (result == null) { serviceDescriptor = result = io.grpc.ServiceDescriptor.newBuilder(SERVICE_NAME) .setSchemaDescriptor(new PredictionServiceFileDescriptorSupplier()) .addMethod(getPredictMethod()) .addMethod(getBatchPredictMethod()) .build(); } } } return result; } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy