Class: Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client

Inherits:
Object
  • Object
show all
Includes:
Paths
Defined in:
lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb

Overview

REST client for the PredictionService service.

A service for online predictions and explanations.

Defined Under Namespace

Classes: Configuration

Instance Attribute Summary collapse

Class Method Summary collapse

Instance Method Summary collapse

Methods included from Paths

#endpoint_path

Constructor Details

#initialize {|config| ... } ⇒ Client

Create a new PredictionService REST client object.

Examples:


# Create a client using the default configuration
client = ::Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a client using a custom configuration
client = ::Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new do |config|
  config.timeout = 10.0
end

Yields:

  • (config)

    Configure the PredictionService client.

Yield Parameters:



128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 128

def initialize
  # Create the configuration object
  @config = Configuration.new Client.configure

  # Yield the configuration if needed
  yield @config if block_given?

  # Create credentials
  credentials = @config.credentials
  # Use self-signed JWT if the endpoint is unchanged from default,
  # but only if the default endpoint does not have a region prefix.
  enable_self_signed_jwt = @config.endpoint.nil? ||
                           (@config.endpoint == Configuration::DEFAULT_ENDPOINT &&
                           !@config.endpoint.split(".").first.include?("-"))
  credentials ||= Credentials.default scope: @config.scope,
                                      enable_self_signed_jwt: enable_self_signed_jwt
  if credentials.is_a?(::String) || credentials.is_a?(::Hash)
    credentials = Credentials.new credentials, scope: @config.scope
  end

  @quota_project_id = @config.quota_project
  @quota_project_id ||= credentials.quota_project_id if credentials.respond_to? :quota_project_id

  @prediction_service_stub = ::Google::Cloud::AIPlatform::V1::PredictionService::Rest::ServiceStub.new(
    endpoint: @config.endpoint,
    endpoint_template: DEFAULT_ENDPOINT_TEMPLATE,
    universe_domain: @config.universe_domain,
    credentials: credentials
  )

  @location_client = Google::Cloud::Location::Locations::Rest::Client.new do |config|
    config.credentials = credentials
    config.quota_project = @quota_project_id
    config.endpoint = @prediction_service_stub.endpoint
    config.universe_domain = @prediction_service_stub.universe_domain
    config.bindings_override = @config.bindings_override
  end

  @iam_policy_client = Google::Iam::V1::IAMPolicy::Rest::Client.new do |config|
    config.credentials = credentials
    config.quota_project = @quota_project_id
    config.endpoint = @prediction_service_stub.endpoint
    config.universe_domain = @prediction_service_stub.universe_domain
    config.bindings_override = @config.bindings_override
  end
end

Instance Attribute Details

#iam_policy_clientGoogle::Iam::V1::IAMPolicy::Rest::Client (readonly)

Get the associated client for mix-in of the IAMPolicy.

Returns:

  • (Google::Iam::V1::IAMPolicy::Rest::Client)


187
188
189
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 187

def iam_policy_client
  @iam_policy_client
end

#location_clientGoogle::Cloud::Location::Locations::Rest::Client (readonly)

Get the associated client for mix-in of the Locations.

Returns:

  • (Google::Cloud::Location::Locations::Rest::Client)


180
181
182
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 180

def location_client
  @location_client
end

Class Method Details

.configure {|config| ... } ⇒ Client::Configuration

Configure the PredictionService Client class.

See Configuration for a description of the configuration fields.

Examples:


# Modify the configuration for all PredictionService clients
::Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.configure do |config|
  config.timeout = 10.0
end

Yields:

  • (config)

    Configure the Client client.

Yield Parameters:

Returns:



66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 66

def self.configure
  @configure ||= begin
    namespace = ["Google", "Cloud", "AIPlatform", "V1"]
    parent_config = while namespace.any?
                      parent_name = namespace.join "::"
                      parent_const = const_get parent_name
                      break parent_const.configure if parent_const.respond_to? :configure
                      namespace.pop
                    end
    default_config = Client::Configuration.new parent_config

    default_config
  end
  yield @configure if block_given?
  @configure
end

Instance Method Details

#configure {|config| ... } ⇒ Client::Configuration

Configure the PredictionService Client instance.

The configuration is set to the derived mode, meaning that values can be changed, but structural changes (adding new fields, etc.) are not allowed. Structural changes should be made on configure.

See Configuration for a description of the configuration fields.

Yields:

  • (config)

    Configure the Client client.

Yield Parameters:

Returns:



98
99
100
101
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 98

def configure
  yield @config if block_given?
  @config
end

#direct_predict(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::DirectPredictResponse #direct_predict(endpoint: nil, inputs: nil, parameters: nil) ⇒ ::Google::Cloud::AIPlatform::V1::DirectPredictResponse

Perform an unary online prediction request to a gRPC model server for Vertex first-party products and frameworks.

Examples:

Basic example

require "google/cloud/ai_platform/v1"

# Create a client object. The client can be reused for multiple calls.
client = Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a request. To set request fields, pass in keyword arguments.
request = Google::Cloud::AIPlatform::V1::DirectPredictRequest.new

# Call the direct_predict method.
result = client.direct_predict request

# The returned object is of type Google::Cloud::AIPlatform::V1::DirectPredictResponse.
p result

Overloads:

  • #direct_predict(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::DirectPredictResponse

    Pass arguments to direct_predict via a request object, either of type DirectPredictRequest or an equivalent Hash.

    Parameters:

    • request (::Google::Cloud::AIPlatform::V1::DirectPredictRequest, ::Hash)

      A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.

    • options (::Gapic::CallOptions, ::Hash) (defaults to: nil)

      Overrides the default settings for this call, e.g, timeout, retries etc. Optional.

  • #direct_predict(endpoint: nil, inputs: nil, parameters: nil) ⇒ ::Google::Cloud::AIPlatform::V1::DirectPredictResponse

    Pass arguments to direct_predict via keyword arguments. Note that at least one keyword argument is required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash as a request object (see above).

    Parameters:

    • endpoint (::String) (defaults to: nil)

      Required. The name of the Endpoint requested to serve the prediction. Format: projects/{project}/locations/{location}/endpoints/{endpoint}

    • inputs (::Array<::Google::Cloud::AIPlatform::V1::Tensor, ::Hash>) (defaults to: nil)

      The prediction input.

    • parameters (::Google::Cloud::AIPlatform::V1::Tensor, ::Hash) (defaults to: nil)

      The parameters that govern the prediction.

Yields:

  • (result, operation)

    Access the result along with the TransportOperation object

Yield Parameters:

Returns:

Raises:

  • (::Google::Cloud::Error)

    if the REST call is aborted.



532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 532

def direct_predict request, options = nil
  raise ::ArgumentError, "request must be provided" if request.nil?

  request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::AIPlatform::V1::DirectPredictRequest

  # Converts hash and nil to an options object
  options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h

  # Customize the options with defaults
   = @config.rpcs.direct_predict..to_h

  # Set x-goog-api-client, x-goog-user-project and x-goog-api-version headers
  [:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
    lib_name: @config.lib_name, lib_version: @config.lib_version,
    gapic_version: ::Google::Cloud::AIPlatform::V1::VERSION,
    transports_version_send: [:rest]

  [:"x-goog-api-version"] = API_VERSION unless API_VERSION.empty?
  [:"x-goog-user-project"] = @quota_project_id if @quota_project_id

  options.apply_defaults timeout:      @config.rpcs.direct_predict.timeout,
                         metadata:     ,
                         retry_policy: @config.rpcs.direct_predict.retry_policy

  options.apply_defaults timeout:      @config.timeout,
                         metadata:     @config.,
                         retry_policy: @config.retry_policy

  @prediction_service_stub.direct_predict request, options do |result, operation|
    yield result, operation if block_given?
    return result
  end
rescue ::Gapic::Rest::Error => e
  raise ::Google::Cloud::Error.from_error(e)
end

#direct_raw_predict(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::DirectRawPredictResponse #direct_raw_predict(endpoint: nil, method_name: nil, input: nil) ⇒ ::Google::Cloud::AIPlatform::V1::DirectRawPredictResponse

Perform an unary online prediction request to a gRPC model server for custom containers.

Examples:

Basic example

require "google/cloud/ai_platform/v1"

# Create a client object. The client can be reused for multiple calls.
client = Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a request. To set request fields, pass in keyword arguments.
request = Google::Cloud::AIPlatform::V1::DirectRawPredictRequest.new

# Call the direct_raw_predict method.
result = client.direct_raw_predict request

# The returned object is of type Google::Cloud::AIPlatform::V1::DirectRawPredictResponse.
p result

Overloads:

  • #direct_raw_predict(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::DirectRawPredictResponse

    Pass arguments to direct_raw_predict via a request object, either of type DirectRawPredictRequest or an equivalent Hash.

    Parameters:

    • request (::Google::Cloud::AIPlatform::V1::DirectRawPredictRequest, ::Hash)

      A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.

    • options (::Gapic::CallOptions, ::Hash) (defaults to: nil)

      Overrides the default settings for this call, e.g, timeout, retries etc. Optional.

  • #direct_raw_predict(endpoint: nil, method_name: nil, input: nil) ⇒ ::Google::Cloud::AIPlatform::V1::DirectRawPredictResponse

    Pass arguments to direct_raw_predict via keyword arguments. Note that at least one keyword argument is required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash as a request object (see above).

    Parameters:

    • endpoint (::String) (defaults to: nil)

      Required. The name of the Endpoint requested to serve the prediction. Format: projects/{project}/locations/{location}/endpoints/{endpoint}

    • method_name (::String) (defaults to: nil)

      Fully qualified name of the API method being invoked to perform predictions.

      Format: /namespace.Service/Method/ Example: /tensorflow.serving.PredictionService/Predict

    • input (::String) (defaults to: nil)

      The prediction input.

Yields:

  • (result, operation)

    Access the result along with the TransportOperation object

Yield Parameters:

Returns:

Raises:

  • (::Google::Cloud::Error)

    if the REST call is aborted.



624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 624

def direct_raw_predict request, options = nil
  raise ::ArgumentError, "request must be provided" if request.nil?

  request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::AIPlatform::V1::DirectRawPredictRequest

  # Converts hash and nil to an options object
  options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h

  # Customize the options with defaults
   = @config.rpcs.direct_raw_predict..to_h

  # Set x-goog-api-client, x-goog-user-project and x-goog-api-version headers
  [:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
    lib_name: @config.lib_name, lib_version: @config.lib_version,
    gapic_version: ::Google::Cloud::AIPlatform::V1::VERSION,
    transports_version_send: [:rest]

  [:"x-goog-api-version"] = API_VERSION unless API_VERSION.empty?
  [:"x-goog-user-project"] = @quota_project_id if @quota_project_id

  options.apply_defaults timeout:      @config.rpcs.direct_raw_predict.timeout,
                         metadata:     ,
                         retry_policy: @config.rpcs.direct_raw_predict.retry_policy

  options.apply_defaults timeout:      @config.timeout,
                         metadata:     @config.,
                         retry_policy: @config.retry_policy

  @prediction_service_stub.direct_raw_predict request, options do |result, operation|
    yield result, operation if block_given?
    return result
  end
rescue ::Gapic::Rest::Error => e
  raise ::Google::Cloud::Error.from_error(e)
end

#explain(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::ExplainResponse #explain(endpoint: nil, instances: nil, parameters: nil, explanation_spec_override: nil, deployed_model_id: nil) ⇒ ::Google::Cloud::AIPlatform::V1::ExplainResponse

Perform an online explanation.

If deployed_model_id is specified, the corresponding DeployModel must have explanation_spec populated. If deployed_model_id is not specified, all DeployedModels must have explanation_spec populated.

Examples:

Basic example

require "google/cloud/ai_platform/v1"

# Create a client object. The client can be reused for multiple calls.
client = Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a request. To set request fields, pass in keyword arguments.
request = Google::Cloud::AIPlatform::V1::ExplainRequest.new

# Call the explain method.
result = client.explain request

# The returned object is of type Google::Cloud::AIPlatform::V1::ExplainResponse.
p result

Overloads:

  • #explain(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::ExplainResponse

    Pass arguments to explain via a request object, either of type ExplainRequest or an equivalent Hash.

    Parameters:

    • request (::Google::Cloud::AIPlatform::V1::ExplainRequest, ::Hash)

      A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.

    • options (::Gapic::CallOptions, ::Hash) (defaults to: nil)

      Overrides the default settings for this call, e.g, timeout, retries etc. Optional.

  • #explain(endpoint: nil, instances: nil, parameters: nil, explanation_spec_override: nil, deployed_model_id: nil) ⇒ ::Google::Cloud::AIPlatform::V1::ExplainResponse

    Pass arguments to explain via keyword arguments. Note that at least one keyword argument is required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash as a request object (see above).

    Parameters:

    • endpoint (::String) (defaults to: nil)

      Required. The name of the Endpoint requested to serve the explanation. Format: projects/{project}/locations/{location}/endpoints/{endpoint}

    • instances (::Array<::Google::Protobuf::Value, ::Hash>) (defaults to: nil)

      Required. The instances that are the input to the explanation call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the explanation call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' [Model's][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] instance_schema_uri.

    • parameters (::Google::Protobuf::Value, ::Hash) (defaults to: nil)

      The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' [Model's ][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] parameters_schema_uri.

    • explanation_spec_override (::Google::Cloud::AIPlatform::V1::ExplanationSpecOverride, ::Hash) (defaults to: nil)

      If specified, overrides the explanation_spec of the DeployedModel. Can be used for explaining prediction results with different configurations, such as:

      • Explaining top-5 predictions results as opposed to top-1;
      • Increasing path count or step count of the attribution methods to reduce approximate errors;
      • Using different baselines for explaining the prediction results.
    • deployed_model_id (::String) (defaults to: nil)

      If specified, this ExplainRequest will be served by the chosen DeployedModel, overriding Endpoint.traffic_split.

Yields:

  • (result, operation)

    Access the result along with the TransportOperation object

Yield Parameters:

Returns:

Raises:

  • (::Google::Cloud::Error)

    if the REST call is aborted.



834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 834

def explain request, options = nil
  raise ::ArgumentError, "request must be provided" if request.nil?

  request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::AIPlatform::V1::ExplainRequest

  # Converts hash and nil to an options object
  options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h

  # Customize the options with defaults
   = @config.rpcs.explain..to_h

  # Set x-goog-api-client, x-goog-user-project and x-goog-api-version headers
  [:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
    lib_name: @config.lib_name, lib_version: @config.lib_version,
    gapic_version: ::Google::Cloud::AIPlatform::V1::VERSION,
    transports_version_send: [:rest]

  [:"x-goog-api-version"] = API_VERSION unless API_VERSION.empty?
  [:"x-goog-user-project"] = @quota_project_id if @quota_project_id

  options.apply_defaults timeout:      @config.rpcs.explain.timeout,
                         metadata:     ,
                         retry_policy: @config.rpcs.explain.retry_policy

  options.apply_defaults timeout:      @config.timeout,
                         metadata:     @config.,
                         retry_policy: @config.retry_policy

  @prediction_service_stub.explain request, options do |result, operation|
    yield result, operation if block_given?
    return result
  end
rescue ::Gapic::Rest::Error => e
  raise ::Google::Cloud::Error.from_error(e)
end

#generate_content(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::GenerateContentResponse #generate_content(model: nil, contents: nil, system_instruction: nil, tools: nil, tool_config: nil, labels: nil, safety_settings: nil, generation_config: nil) ⇒ ::Google::Cloud::AIPlatform::V1::GenerateContentResponse

Generate content with multimodal inputs.

Examples:

Basic example

require "google/cloud/ai_platform/v1"

# Create a client object. The client can be reused for multiple calls.
client = Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a request. To set request fields, pass in keyword arguments.
request = Google::Cloud::AIPlatform::V1::GenerateContentRequest.new

# Call the generate_content method.
result = client.generate_content request

# The returned object is of type Google::Cloud::AIPlatform::V1::GenerateContentResponse.
p result

Overloads:

  • #generate_content(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::GenerateContentResponse

    Pass arguments to generate_content via a request object, either of type GenerateContentRequest or an equivalent Hash.

    Parameters:

    • request (::Google::Cloud::AIPlatform::V1::GenerateContentRequest, ::Hash)

      A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.

    • options (::Gapic::CallOptions, ::Hash) (defaults to: nil)

      Overrides the default settings for this call, e.g, timeout, retries etc. Optional.

  • #generate_content(model: nil, contents: nil, system_instruction: nil, tools: nil, tool_config: nil, labels: nil, safety_settings: nil, generation_config: nil) ⇒ ::Google::Cloud::AIPlatform::V1::GenerateContentResponse

    Pass arguments to generate_content via keyword arguments. Note that at least one keyword argument is required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash as a request object (see above).

    Parameters:

    • model (::String) (defaults to: nil)

      Required. The fully qualified name of the publisher model or tuned model endpoint to use.

      Publisher model format: projects/{project}/locations/{location}/publishers/*/models/*

      Tuned model endpoint format: projects/{project}/locations/{location}/endpoints/{endpoint}

    • contents (::Array<::Google::Cloud::AIPlatform::V1::Content, ::Hash>) (defaults to: nil)

      Required. The content of the current conversation with the model.

      For single-turn queries, this is a single instance. For multi-turn queries, this is a repeated field that contains conversation history + latest request.

    • system_instruction (::Google::Cloud::AIPlatform::V1::Content, ::Hash) (defaults to: nil)

      Optional. The user provided system instructions for the model. Note: only text should be used in parts and content in each part will be in a separate paragraph.

    • tools (::Array<::Google::Cloud::AIPlatform::V1::Tool, ::Hash>) (defaults to: nil)

      Optional. A list of Tools the model may use to generate the next response.

      A Tool is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model.

    • tool_config (::Google::Cloud::AIPlatform::V1::ToolConfig, ::Hash) (defaults to: nil)

      Optional. Tool config. This config is shared for all tools provided in the request.

    • labels (::Hash{::String => ::String}) (defaults to: nil)

      Optional. The labels with user-defined metadata for the request. It is used for billing and reporting only.

      Label keys and values can be no longer than 63 characters (Unicode codepoints) and can only contain lowercase letters, numeric characters, underscores, and dashes. International characters are allowed. Label values are optional. Label keys must start with a letter.

    • safety_settings (::Array<::Google::Cloud::AIPlatform::V1::SafetySetting, ::Hash>) (defaults to: nil)

      Optional. Per request settings for blocking unsafe content. Enforced on GenerateContentResponse.candidates.

    • generation_config (::Google::Cloud::AIPlatform::V1::GenerationConfig, ::Hash) (defaults to: nil)

      Optional. Generation config.

Yields:

  • (result, operation)

    Access the result along with the TransportOperation object

Yield Parameters:

Returns:

Raises:

  • (::Google::Cloud::Error)

    if the REST call is aborted.



953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 953

def generate_content request, options = nil
  raise ::ArgumentError, "request must be provided" if request.nil?

  request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::AIPlatform::V1::GenerateContentRequest

  # Converts hash and nil to an options object
  options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h

  # Customize the options with defaults
   = @config.rpcs.generate_content..to_h

  # Set x-goog-api-client, x-goog-user-project and x-goog-api-version headers
  [:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
    lib_name: @config.lib_name, lib_version: @config.lib_version,
    gapic_version: ::Google::Cloud::AIPlatform::V1::VERSION,
    transports_version_send: [:rest]

  [:"x-goog-api-version"] = API_VERSION unless API_VERSION.empty?
  [:"x-goog-user-project"] = @quota_project_id if @quota_project_id

  options.apply_defaults timeout:      @config.rpcs.generate_content.timeout,
                         metadata:     ,
                         retry_policy: @config.rpcs.generate_content.retry_policy

  options.apply_defaults timeout:      @config.timeout,
                         metadata:     @config.,
                         retry_policy: @config.retry_policy

  @prediction_service_stub.generate_content request, options do |result, operation|
    yield result, operation if block_given?
    return result
  end
rescue ::Gapic::Rest::Error => e
  raise ::Google::Cloud::Error.from_error(e)
end

#predict(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::PredictResponse #predict(endpoint: nil, instances: nil, parameters: nil) ⇒ ::Google::Cloud::AIPlatform::V1::PredictResponse

Perform an online prediction.

Examples:

Basic example

require "google/cloud/ai_platform/v1"

# Create a client object. The client can be reused for multiple calls.
client = Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a request. To set request fields, pass in keyword arguments.
request = Google::Cloud::AIPlatform::V1::PredictRequest.new

# Call the predict method.
result = client.predict request

# The returned object is of type Google::Cloud::AIPlatform::V1::PredictResponse.
p result

Overloads:

  • #predict(request, options = nil) ⇒ ::Google::Cloud::AIPlatform::V1::PredictResponse

    Pass arguments to predict via a request object, either of type Google::Cloud::AIPlatform::V1::PredictRequest or an equivalent Hash.

    Parameters:

    • request (::Google::Cloud::AIPlatform::V1::PredictRequest, ::Hash)

      A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.

    • options (::Gapic::CallOptions, ::Hash) (defaults to: nil)

      Overrides the default settings for this call, e.g, timeout, retries etc. Optional.

  • #predict(endpoint: nil, instances: nil, parameters: nil) ⇒ ::Google::Cloud::AIPlatform::V1::PredictResponse

    Pass arguments to predict via keyword arguments. Note that at least one keyword argument is required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash as a request object (see above).

    Parameters:

    • endpoint (::String) (defaults to: nil)

      Required. The name of the Endpoint requested to serve the prediction. Format: projects/{project}/locations/{location}/endpoints/{endpoint}

    • instances (::Array<::Google::Protobuf::Value, ::Hash>) (defaults to: nil)

      Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' [Model's][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] instance_schema_uri.

    • parameters (::Google::Protobuf::Value, ::Hash) (defaults to: nil)

      The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' [Model's ][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] parameters_schema_uri.

Yields:

  • (result, operation)

    Access the result along with the TransportOperation object

Yield Parameters:

Returns:

Raises:

  • (::Google::Cloud::Error)

    if the REST call is aborted.



252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 252

def predict request, options = nil
  raise ::ArgumentError, "request must be provided" if request.nil?

  request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::AIPlatform::V1::PredictRequest

  # Converts hash and nil to an options object
  options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h

  # Customize the options with defaults
   = @config.rpcs.predict..to_h

  # Set x-goog-api-client, x-goog-user-project and x-goog-api-version headers
  [:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
    lib_name: @config.lib_name, lib_version: @config.lib_version,
    gapic_version: ::Google::Cloud::AIPlatform::V1::VERSION,
    transports_version_send: [:rest]

  [:"x-goog-api-version"] = API_VERSION unless API_VERSION.empty?
  [:"x-goog-user-project"] = @quota_project_id if @quota_project_id

  options.apply_defaults timeout:      @config.rpcs.predict.timeout,
                         metadata:     ,
                         retry_policy: @config.rpcs.predict.retry_policy

  options.apply_defaults timeout:      @config.timeout,
                         metadata:     @config.,
                         retry_policy: @config.retry_policy

  @prediction_service_stub.predict request, options do |result, operation|
    yield result, operation if block_given?
    return result
  end
rescue ::Gapic::Rest::Error => e
  raise ::Google::Cloud::Error.from_error(e)
end

#raw_predict(request, options = nil) ⇒ ::Google::Api::HttpBody #raw_predict(endpoint: nil, http_body: nil) ⇒ ::Google::Api::HttpBody

Perform an online prediction with an arbitrary HTTP payload.

The response includes the following HTTP headers:

  • X-Vertex-AI-Endpoint-Id: ID of the Endpoint that served this prediction.

  • X-Vertex-AI-Deployed-Model-Id: ID of the Endpoint's DeployedModel that served this prediction.

Examples:

Basic example

require "google/cloud/ai_platform/v1"

# Create a client object. The client can be reused for multiple calls.
client = Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a request. To set request fields, pass in keyword arguments.
request = Google::Cloud::AIPlatform::V1::RawPredictRequest.new

# Call the raw_predict method.
result = client.raw_predict request

# The returned object is of type Google::Api::HttpBody.
p result

Overloads:

  • #raw_predict(request, options = nil) ⇒ ::Google::Api::HttpBody

    Pass arguments to raw_predict via a request object, either of type RawPredictRequest or an equivalent Hash.

    Parameters:

    • request (::Google::Cloud::AIPlatform::V1::RawPredictRequest, ::Hash)

      A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.

    • options (::Gapic::CallOptions, ::Hash) (defaults to: nil)

      Overrides the default settings for this call, e.g, timeout, retries etc. Optional.

  • #raw_predict(endpoint: nil, http_body: nil) ⇒ ::Google::Api::HttpBody

    Pass arguments to raw_predict via keyword arguments. Note that at least one keyword argument is required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash as a request object (see above).

    Parameters:

    • endpoint (::String) (defaults to: nil)

      Required. The name of the Endpoint requested to serve the prediction. Format: projects/{project}/locations/{location}/endpoints/{endpoint}

    • http_body (::Google::Api::HttpBody, ::Hash) (defaults to: nil)

      The prediction input. Supports HTTP headers and arbitrary data payload.

      A DeployedModel may have an upper limit on the number of instances it supports per request. When this limit it is exceeded for an AutoML model, the RawPredict method returns an error. When this limit is exceeded for a custom-trained model, the behavior varies depending on the model.

      You can specify the schema for each instance in the predict_schemata.instance_schema_uri field when you create a Model. This schema applies when you deploy the Model as a DeployedModel to an Endpoint and use the RawPredict method.

Yields:

  • (result, operation)

    Access the result along with the TransportOperation object

Yield Parameters:

Returns:

Raises:

  • (::Google::Cloud::Error)

    if the REST call is aborted.



359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 359

def raw_predict request, options = nil
  raise ::ArgumentError, "request must be provided" if request.nil?

  request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::AIPlatform::V1::RawPredictRequest

  # Converts hash and nil to an options object
  options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h

  # Customize the options with defaults
   = @config.rpcs.raw_predict..to_h

  # Set x-goog-api-client, x-goog-user-project and x-goog-api-version headers
  [:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
    lib_name: @config.lib_name, lib_version: @config.lib_version,
    gapic_version: ::Google::Cloud::AIPlatform::V1::VERSION,
    transports_version_send: [:rest]

  [:"x-goog-api-version"] = API_VERSION unless API_VERSION.empty?
  [:"x-goog-user-project"] = @quota_project_id if @quota_project_id

  options.apply_defaults timeout:      @config.rpcs.raw_predict.timeout,
                         metadata:     ,
                         retry_policy: @config.rpcs.raw_predict.retry_policy

  options.apply_defaults timeout:      @config.timeout,
                         metadata:     @config.,
                         retry_policy: @config.retry_policy

  @prediction_service_stub.raw_predict request, options do |result, operation|
    yield result, operation if block_given?
    return result
  end
rescue ::Gapic::Rest::Error => e
  raise ::Google::Cloud::Error.from_error(e)
end

#server_streaming_predict(request, options = nil) ⇒ ::Enumerable<::Google::Cloud::AIPlatform::V1::StreamingPredictResponse> #server_streaming_predict(endpoint: nil, inputs: nil, parameters: nil) ⇒ ::Enumerable<::Google::Cloud::AIPlatform::V1::StreamingPredictResponse>

Perform a server-side streaming online prediction request for Vertex LLM streaming.

Examples:

Basic example

require "google/cloud/ai_platform/v1"

# Create a client object. The client can be reused for multiple calls.
client = Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a request. To set request fields, pass in keyword arguments.
request = Google::Cloud::AIPlatform::V1::StreamingPredictRequest.new

# Call the server_streaming_predict method to start streaming.
output = client.server_streaming_predict request

# The returned object is a streamed enumerable yielding elements of type
# ::Google::Cloud::AIPlatform::V1::StreamingPredictResponse
output.each do |current_response|
  p current_response
end

Overloads:

  • #server_streaming_predict(request, options = nil) ⇒ ::Enumerable<::Google::Cloud::AIPlatform::V1::StreamingPredictResponse>

    Pass arguments to server_streaming_predict via a request object, either of type StreamingPredictRequest or an equivalent Hash.

    Parameters:

    • request (::Google::Cloud::AIPlatform::V1::StreamingPredictRequest, ::Hash)

      A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.

    • options (::Gapic::CallOptions, ::Hash) (defaults to: nil)

      Overrides the default settings for this call, e.g, timeout, retries etc. Optional.

  • #server_streaming_predict(endpoint: nil, inputs: nil, parameters: nil) ⇒ ::Enumerable<::Google::Cloud::AIPlatform::V1::StreamingPredictResponse>

    Pass arguments to server_streaming_predict via keyword arguments. Note that at least one keyword argument is required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash as a request object (see above).

    Parameters:

    • endpoint (::String) (defaults to: nil)

      Required. The name of the Endpoint requested to serve the prediction. Format: projects/{project}/locations/{location}/endpoints/{endpoint}

    • inputs (::Array<::Google::Cloud::AIPlatform::V1::Tensor, ::Hash>) (defaults to: nil)

      The prediction input.

    • parameters (::Google::Cloud::AIPlatform::V1::Tensor, ::Hash) (defaults to: nil)

      The parameters that govern the prediction.

Returns:

Raises:

  • (::Google::Cloud::Error)

    if the REST call is aborted.



709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 709

def server_streaming_predict request, options = nil
  raise ::ArgumentError, "request must be provided" if request.nil?

  request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::AIPlatform::V1::StreamingPredictRequest

  # Converts hash and nil to an options object
  options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h

  # Customize the options with defaults
   = @config.rpcs.server_streaming_predict..to_h

  # Set x-goog-api-client, x-goog-user-project and x-goog-api-version headers
  [:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
    lib_name: @config.lib_name, lib_version: @config.lib_version,
    gapic_version: ::Google::Cloud::AIPlatform::V1::VERSION,
    transports_version_send: [:rest]

  [:"x-goog-api-version"] = API_VERSION unless API_VERSION.empty?
  [:"x-goog-user-project"] = @quota_project_id if @quota_project_id

  options.apply_defaults timeout:      @config.rpcs.server_streaming_predict.timeout,
                         metadata:     ,
                         retry_policy: @config.rpcs.server_streaming_predict.retry_policy

  options.apply_defaults timeout:      @config.timeout,
                         metadata:     @config.,
                         retry_policy: @config.retry_policy

  ::Gapic::Rest::ServerStream.new(
    ::Google::Cloud::AIPlatform::V1::StreamingPredictResponse,
    ::Gapic::Rest::ThreadedEnumerator.new do |in_q, out_q|
      @prediction_service_stub.server_streaming_predict request, options do |chunk|
        in_q.deq
        out_q.enq chunk
      end
    end
  )
rescue ::Gapic::Rest::Error => e
  raise ::Google::Cloud::Error.from_error(e)
end

#stream_generate_content(request, options = nil) ⇒ ::Enumerable<::Google::Cloud::AIPlatform::V1::GenerateContentResponse> #stream_generate_content(model: nil, contents: nil, system_instruction: nil, tools: nil, tool_config: nil, labels: nil, safety_settings: nil, generation_config: nil) ⇒ ::Enumerable<::Google::Cloud::AIPlatform::V1::GenerateContentResponse>

Generate content with multimodal inputs with streaming support.

Examples:

Basic example

require "google/cloud/ai_platform/v1"

# Create a client object. The client can be reused for multiple calls.
client = Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a request. To set request fields, pass in keyword arguments.
request = Google::Cloud::AIPlatform::V1::GenerateContentRequest.new

# Call the stream_generate_content method to start streaming.
output = client.stream_generate_content request

# The returned object is a streamed enumerable yielding elements of type
# ::Google::Cloud::AIPlatform::V1::GenerateContentResponse
output.each do |current_response|
  p current_response
end

Overloads:

  • #stream_generate_content(request, options = nil) ⇒ ::Enumerable<::Google::Cloud::AIPlatform::V1::GenerateContentResponse>

    Pass arguments to stream_generate_content via a request object, either of type GenerateContentRequest or an equivalent Hash.

    Parameters:

    • request (::Google::Cloud::AIPlatform::V1::GenerateContentRequest, ::Hash)

      A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.

    • options (::Gapic::CallOptions, ::Hash) (defaults to: nil)

      Overrides the default settings for this call, e.g, timeout, retries etc. Optional.

  • #stream_generate_content(model: nil, contents: nil, system_instruction: nil, tools: nil, tool_config: nil, labels: nil, safety_settings: nil, generation_config: nil) ⇒ ::Enumerable<::Google::Cloud::AIPlatform::V1::GenerateContentResponse>

    Pass arguments to stream_generate_content via keyword arguments. Note that at least one keyword argument is required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash as a request object (see above).

    Parameters:

    • model (::String) (defaults to: nil)

      Required. The fully qualified name of the publisher model or tuned model endpoint to use.

      Publisher model format: projects/{project}/locations/{location}/publishers/*/models/*

      Tuned model endpoint format: projects/{project}/locations/{location}/endpoints/{endpoint}

    • contents (::Array<::Google::Cloud::AIPlatform::V1::Content, ::Hash>) (defaults to: nil)

      Required. The content of the current conversation with the model.

      For single-turn queries, this is a single instance. For multi-turn queries, this is a repeated field that contains conversation history + latest request.

    • system_instruction (::Google::Cloud::AIPlatform::V1::Content, ::Hash) (defaults to: nil)

      Optional. The user provided system instructions for the model. Note: only text should be used in parts and content in each part will be in a separate paragraph.

    • tools (::Array<::Google::Cloud::AIPlatform::V1::Tool, ::Hash>) (defaults to: nil)

      Optional. A list of Tools the model may use to generate the next response.

      A Tool is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model.

    • tool_config (::Google::Cloud::AIPlatform::V1::ToolConfig, ::Hash) (defaults to: nil)

      Optional. Tool config. This config is shared for all tools provided in the request.

    • labels (::Hash{::String => ::String}) (defaults to: nil)

      Optional. The labels with user-defined metadata for the request. It is used for billing and reporting only.

      Label keys and values can be no longer than 63 characters (Unicode codepoints) and can only contain lowercase letters, numeric characters, underscores, and dashes. International characters are allowed. Label values are optional. Label keys must start with a letter.

    • safety_settings (::Array<::Google::Cloud::AIPlatform::V1::SafetySetting, ::Hash>) (defaults to: nil)

      Optional. Per request settings for blocking unsafe content. Enforced on GenerateContentResponse.candidates.

    • generation_config (::Google::Cloud::AIPlatform::V1::GenerationConfig, ::Hash) (defaults to: nil)

      Optional. Generation config.

Returns:

Raises:

  • (::Google::Cloud::Error)

    if the REST call is aborted.



1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 1071

def stream_generate_content request, options = nil
  raise ::ArgumentError, "request must be provided" if request.nil?

  request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::AIPlatform::V1::GenerateContentRequest

  # Converts hash and nil to an options object
  options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h

  # Customize the options with defaults
   = @config.rpcs.stream_generate_content..to_h

  # Set x-goog-api-client, x-goog-user-project and x-goog-api-version headers
  [:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
    lib_name: @config.lib_name, lib_version: @config.lib_version,
    gapic_version: ::Google::Cloud::AIPlatform::V1::VERSION,
    transports_version_send: [:rest]

  [:"x-goog-api-version"] = API_VERSION unless API_VERSION.empty?
  [:"x-goog-user-project"] = @quota_project_id if @quota_project_id

  options.apply_defaults timeout:      @config.rpcs.stream_generate_content.timeout,
                         metadata:     ,
                         retry_policy: @config.rpcs.stream_generate_content.retry_policy

  options.apply_defaults timeout:      @config.timeout,
                         metadata:     @config.,
                         retry_policy: @config.retry_policy

  ::Gapic::Rest::ServerStream.new(
    ::Google::Cloud::AIPlatform::V1::GenerateContentResponse,
    ::Gapic::Rest::ThreadedEnumerator.new do |in_q, out_q|
      @prediction_service_stub.stream_generate_content request, options do |chunk|
        in_q.deq
        out_q.enq chunk
      end
    end
  )
rescue ::Gapic::Rest::Error => e
  raise ::Google::Cloud::Error.from_error(e)
end

#stream_raw_predict(request, options = nil) ⇒ ::Enumerable<::Google::Api::HttpBody> #stream_raw_predict(endpoint: nil, http_body: nil) ⇒ ::Enumerable<::Google::Api::HttpBody>

Perform a streaming online prediction with an arbitrary HTTP payload.

Examples:

Basic example

require "google/cloud/ai_platform/v1"

# Create a client object. The client can be reused for multiple calls.
client = Google::Cloud::AIPlatform::V1::PredictionService::Rest::Client.new

# Create a request. To set request fields, pass in keyword arguments.
request = Google::Cloud::AIPlatform::V1::StreamRawPredictRequest.new

# Call the stream_raw_predict method to start streaming.
output = client.stream_raw_predict request

# The returned object is a streamed enumerable yielding elements of type
# ::Google::Api::HttpBody
output.each do |current_response|
  p current_response
end

Overloads:

  • #stream_raw_predict(request, options = nil) ⇒ ::Enumerable<::Google::Api::HttpBody>

    Pass arguments to stream_raw_predict via a request object, either of type StreamRawPredictRequest or an equivalent Hash.

    Parameters:

    • request (::Google::Cloud::AIPlatform::V1::StreamRawPredictRequest, ::Hash)

      A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.

    • options (::Gapic::CallOptions, ::Hash) (defaults to: nil)

      Overrides the default settings for this call, e.g, timeout, retries etc. Optional.

  • #stream_raw_predict(endpoint: nil, http_body: nil) ⇒ ::Enumerable<::Google::Api::HttpBody>

    Pass arguments to stream_raw_predict via keyword arguments. Note that at least one keyword argument is required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash as a request object (see above).

    Parameters:

    • endpoint (::String) (defaults to: nil)

      Required. The name of the Endpoint requested to serve the prediction. Format: projects/{project}/locations/{location}/endpoints/{endpoint}

    • http_body (::Google::Api::HttpBody, ::Hash) (defaults to: nil)

      The prediction input. Supports HTTP headers and arbitrary data payload.

Returns:

Raises:

  • (::Google::Cloud::Error)

    if the REST call is aborted.



441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 441

def stream_raw_predict request, options = nil
  raise ::ArgumentError, "request must be provided" if request.nil?

  request = ::Gapic::Protobuf.coerce request, to: ::Google::Cloud::AIPlatform::V1::StreamRawPredictRequest

  # Converts hash and nil to an options object
  options = ::Gapic::CallOptions.new(**options.to_h) if options.respond_to? :to_h

  # Customize the options with defaults
   = @config.rpcs.stream_raw_predict..to_h

  # Set x-goog-api-client, x-goog-user-project and x-goog-api-version headers
  [:"x-goog-api-client"] ||= ::Gapic::Headers.x_goog_api_client \
    lib_name: @config.lib_name, lib_version: @config.lib_version,
    gapic_version: ::Google::Cloud::AIPlatform::V1::VERSION,
    transports_version_send: [:rest]

  [:"x-goog-api-version"] = API_VERSION unless API_VERSION.empty?
  [:"x-goog-user-project"] = @quota_project_id if @quota_project_id

  options.apply_defaults timeout:      @config.rpcs.stream_raw_predict.timeout,
                         metadata:     ,
                         retry_policy: @config.rpcs.stream_raw_predict.retry_policy

  options.apply_defaults timeout:      @config.timeout,
                         metadata:     @config.,
                         retry_policy: @config.retry_policy

  ::Gapic::Rest::ServerStream.new(
    ::Google::Api::HttpBody,
    ::Gapic::Rest::ThreadedEnumerator.new do |in_q, out_q|
      @prediction_service_stub.stream_raw_predict request, options do |chunk|
        in_q.deq
        out_q.enq chunk
      end
    end
  )
rescue ::Gapic::Rest::Error => e
  raise ::Google::Cloud::Error.from_error(e)
end

#universe_domainString

The effective universe domain

Returns:

  • (String)


108
109
110
# File 'lib/google/cloud/ai_platform/v1/prediction_service/rest/client.rb', line 108

def universe_domain
  @prediction_service_stub.universe_domain
end