Class: Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2BatchUpdateBlobsRequestRequest

Inherits:
Object
  • Object
show all
Includes:
Core::Hashable, Core::JsonObjectSupport
Defined in:
generated/google/apis/remotebuildexecution_v2/classes.rb,
generated/google/apis/remotebuildexecution_v2/representations.rb,
generated/google/apis/remotebuildexecution_v2/representations.rb

Overview

A request corresponding to a single blob that the client wants to upload.

Instance Attribute Summary collapse

Instance Method Summary collapse

Methods included from Core::JsonObjectSupport

#to_json

Methods included from Core::Hashable

process_value, #to_h

Constructor Details

#initialize(**args) ⇒ BuildBazelRemoteExecutionV2BatchUpdateBlobsRequestRequest

Returns a new instance of BuildBazelRemoteExecutionV2BatchUpdateBlobsRequestRequest.



562
563
564
# File 'generated/google/apis/remotebuildexecution_v2/classes.rb', line 562

def initialize(**args)
   update!(**args)
end

Instance Attribute Details

#dataString

The raw binary data. Corresponds to the JSON property data NOTE: Values are automatically base64 encoded/decoded in the client library.

Returns:

  • (String)


529
530
531
# File 'generated/google/apis/remotebuildexecution_v2/classes.rb', line 529

def data
  @data
end

#digestGoogle::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest

A content digest. A digest for a given blob consists of the size of the blob and its hash. The hash algorithm to use is defined by the server, but servers SHOULD use SHA-256. The size is considered to be an integral part of the digest and cannot be separated. That is, even if the hash field is correctly specified but size_bytes is not, the server MUST reject the request. The reason for including the size in the digest is as follows: in a great many cases, the server needs to know the size of the blob it is about to work with prior to starting an operation with it, such as flattening Merkle tree structures or streaming it to a worker. Technically, the server could implement a separate metadata store, but this results in a significantly more complicated implementation as opposed to having the client specify the size up-front (or storing the size along with the digest in every message where digests are embedded). This does mean that the API leaks some implementation details of (what we consider to be) a reasonable server implementation, but we consider this to be a worthwhile tradeoff. When a Digest is used to refer to a proto message, it always refers to the message in binary encoded form. To ensure consistent hashing, clients and servers MUST ensure that they serialize messages according to the following rules, even if there are alternate valid encodings for the same message:

  • Fields are serialized in tag order.
  • There are no unknown fields.
  • There are no duplicate fields.
  • Fields are serialized according to the default semantics for their type. Most protocol buffer implementations will always follow these rules when serializing, but care should be taken to avoid shortcuts. For instance, concatenating two messages to merge them may produce duplicate fields. Corresponds to the JSON property digest


560
561
562
# File 'generated/google/apis/remotebuildexecution_v2/classes.rb', line 560

def digest
  @digest
end

Instance Method Details

#update!(**args) ⇒ Object

Update properties of this object



567
568
569
570
# File 'generated/google/apis/remotebuildexecution_v2/classes.rb', line 567

def update!(**args)
  @data = args[:data] if args.key?(:data)
  @digest = args[:digest] if args.key?(:digest)
end