public final class PySparkJob extends GeneratedMessageV3 implements PySparkJobOrBuilder
A Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN.Protobuf type
google.cloud.dataproc.v1.PySparkJob
Modifier and Type | Class and Description |
---|---|
static class |
PySparkJob.Builder
A Dataproc job for running
[Apache
PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
applications on YARN.
|
GeneratedMessageV3.BuilderParent, GeneratedMessageV3.ExtendableBuilder<MessageType extends GeneratedMessageV3.ExtendableMessage,BuilderType extends GeneratedMessageV3.ExtendableBuilder<MessageType,BuilderType>>, GeneratedMessageV3.ExtendableMessage<MessageType extends GeneratedMessageV3.ExtendableMessage>, GeneratedMessageV3.ExtendableMessageOrBuilder<MessageType extends GeneratedMessageV3.ExtendableMessage>, GeneratedMessageV3.FieldAccessorTable, GeneratedMessageV3.UnusedPrivateParameter
AbstractMessageLite.InternalOneOfEnum
Modifier and Type | Field and Description |
---|---|
static int |
ARCHIVE_URIS_FIELD_NUMBER |
static int |
ARGS_FIELD_NUMBER |
static int |
FILE_URIS_FIELD_NUMBER |
static int |
JAR_FILE_URIS_FIELD_NUMBER |
static int |
LOGGING_CONFIG_FIELD_NUMBER |
static int |
MAIN_PYTHON_FILE_URI_FIELD_NUMBER |
static int |
PROPERTIES_FIELD_NUMBER |
static int |
PYTHON_FILE_URIS_FIELD_NUMBER |
alwaysUseFieldBuilders, unknownFields
memoizedSize
memoizedHashCode
canUseUnsafe, computeStringSize, computeStringSizeNoTag, emptyBooleanList, emptyDoubleList, emptyFloatList, emptyIntList, emptyLongList, getAllFields, getDescriptorForType, getField, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, hasField, hasOneof, isStringEmpty, makeExtensionsImmutable, mergeFromAndMakeImmutableInternal, mutableCopy, mutableCopy, mutableCopy, mutableCopy, mutableCopy, newBooleanList, newBuilderForType, newDoubleList, newFloatList, newIntList, newLongList, parseDelimitedWithIOException, parseDelimitedWithIOException, parseUnknownField, parseUnknownFieldProto3, parseWithIOException, parseWithIOException, parseWithIOException, parseWithIOException, serializeBooleanMapTo, serializeIntegerMapTo, serializeLongMapTo, serializeStringMapTo, writeReplace, writeString, writeStringNoTag
findInitializationErrors, getInitializationErrorString, hashBoolean, hashEnum, hashEnumList, hashFields, hashLong, toString
addAll, addAll, checkByteStringIsUtf8, toByteArray, toByteString, writeDelimitedTo, writeTo
clone, finalize, getClass, notify, notifyAll, wait, wait, wait
findInitializationErrors, getAllFields, getDescriptorForType, getField, getInitializationErrorString, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, hasField, hasOneof
toByteArray, toByteString, writeDelimitedTo, writeTo
public static final int MAIN_PYTHON_FILE_URI_FIELD_NUMBER
public static final int ARGS_FIELD_NUMBER
public static final int PYTHON_FILE_URIS_FIELD_NUMBER
public static final int JAR_FILE_URIS_FIELD_NUMBER
public static final int FILE_URIS_FIELD_NUMBER
public static final int ARCHIVE_URIS_FIELD_NUMBER
public static final int PROPERTIES_FIELD_NUMBER
public static final int LOGGING_CONFIG_FIELD_NUMBER
protected Object newInstance(GeneratedMessageV3.UnusedPrivateParameter unused)
newInstance
in class GeneratedMessageV3
public final UnknownFieldSet getUnknownFields()
getUnknownFields
in interface MessageOrBuilder
getUnknownFields
in class GeneratedMessageV3
public static final Descriptors.Descriptor getDescriptor()
protected MapField internalGetMapField(int number)
internalGetMapField
in class GeneratedMessageV3
protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
internalGetFieldAccessorTable
in class GeneratedMessageV3
public String getMainPythonFileUri()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];
getMainPythonFileUri
in interface PySparkJobOrBuilder
public ByteString getMainPythonFileUriBytes()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1 [(.google.api.field_behavior) = REQUIRED];
getMainPythonFileUriBytes
in interface PySparkJobOrBuilder
public ProtocolStringList getArgsList()
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
getArgsList
in interface PySparkJobOrBuilder
public int getArgsCount()
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
getArgsCount
in interface PySparkJobOrBuilder
public String getArgs(int index)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
getArgs
in interface PySparkJobOrBuilder
index
- The index of the element to return.public ByteString getArgsBytes(int index)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2 [(.google.api.field_behavior) = OPTIONAL];
getArgsBytes
in interface PySparkJobOrBuilder
index
- The index of the value to return.public ProtocolStringList getPythonFileUrisList()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
getPythonFileUrisList
in interface PySparkJobOrBuilder
public int getPythonFileUrisCount()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
getPythonFileUrisCount
in interface PySparkJobOrBuilder
public String getPythonFileUris(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
getPythonFileUris
in interface PySparkJobOrBuilder
index
- The index of the element to return.public ByteString getPythonFileUrisBytes(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3 [(.google.api.field_behavior) = OPTIONAL];
getPythonFileUrisBytes
in interface PySparkJobOrBuilder
index
- The index of the value to return.public ProtocolStringList getJarFileUrisList()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
getJarFileUrisList
in interface PySparkJobOrBuilder
public int getJarFileUrisCount()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
getJarFileUrisCount
in interface PySparkJobOrBuilder
public String getJarFileUris(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
getJarFileUris
in interface PySparkJobOrBuilder
index
- The index of the element to return.public ByteString getJarFileUrisBytes(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4 [(.google.api.field_behavior) = OPTIONAL];
getJarFileUrisBytes
in interface PySparkJobOrBuilder
index
- The index of the value to return.public ProtocolStringList getFileUrisList()
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
getFileUrisList
in interface PySparkJobOrBuilder
public int getFileUrisCount()
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
getFileUrisCount
in interface PySparkJobOrBuilder
public String getFileUris(int index)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
getFileUris
in interface PySparkJobOrBuilder
index
- The index of the element to return.public ByteString getFileUrisBytes(int index)
Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
repeated string file_uris = 5 [(.google.api.field_behavior) = OPTIONAL];
getFileUrisBytes
in interface PySparkJobOrBuilder
index
- The index of the value to return.public ProtocolStringList getArchiveUrisList()
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
getArchiveUrisList
in interface PySparkJobOrBuilder
public int getArchiveUrisCount()
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
getArchiveUrisCount
in interface PySparkJobOrBuilder
public String getArchiveUris(int index)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
getArchiveUris
in interface PySparkJobOrBuilder
index
- The index of the element to return.public ByteString getArchiveUrisBytes(int index)
Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6 [(.google.api.field_behavior) = OPTIONAL];
getArchiveUrisBytes
in interface PySparkJobOrBuilder
index
- The index of the value to return.public int getPropertiesCount()
PySparkJobOrBuilder
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
getPropertiesCount
in interface PySparkJobOrBuilder
public boolean containsProperties(String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
containsProperties
in interface PySparkJobOrBuilder
@Deprecated public Map<String,String> getProperties()
getPropertiesMap()
instead.getProperties
in interface PySparkJobOrBuilder
public Map<String,String> getPropertiesMap()
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
getPropertiesMap
in interface PySparkJobOrBuilder
public String getPropertiesOrDefault(String key, String defaultValue)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
getPropertiesOrDefault
in interface PySparkJobOrBuilder
public String getPropertiesOrThrow(String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7 [(.google.api.field_behavior) = OPTIONAL];
getPropertiesOrThrow
in interface PySparkJobOrBuilder
public boolean hasLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
hasLoggingConfig
in interface PySparkJobOrBuilder
public LoggingConfig getLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
getLoggingConfig
in interface PySparkJobOrBuilder
public LoggingConfigOrBuilder getLoggingConfigOrBuilder()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8 [(.google.api.field_behavior) = OPTIONAL];
getLoggingConfigOrBuilder
in interface PySparkJobOrBuilder
public final boolean isInitialized()
isInitialized
in interface MessageLiteOrBuilder
isInitialized
in class GeneratedMessageV3
public void writeTo(CodedOutputStream output) throws IOException
writeTo
in interface MessageLite
writeTo
in class GeneratedMessageV3
IOException
public int getSerializedSize()
getSerializedSize
in interface MessageLite
getSerializedSize
in class GeneratedMessageV3
public boolean equals(Object obj)
equals
in interface Message
equals
in class AbstractMessage
public int hashCode()
hashCode
in interface Message
hashCode
in class AbstractMessage
public static PySparkJob parseFrom(ByteBuffer data) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(ByteBuffer data, ExtensionRegistryLite extensionRegistry) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(ByteString data) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(ByteString data, ExtensionRegistryLite extensionRegistry) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(byte[] data) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(byte[] data, ExtensionRegistryLite extensionRegistry) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(InputStream input) throws IOException
IOException
public static PySparkJob parseFrom(InputStream input, ExtensionRegistryLite extensionRegistry) throws IOException
IOException
public static PySparkJob parseDelimitedFrom(InputStream input) throws IOException
IOException
public static PySparkJob parseDelimitedFrom(InputStream input, ExtensionRegistryLite extensionRegistry) throws IOException
IOException
public static PySparkJob parseFrom(CodedInputStream input) throws IOException
IOException
public static PySparkJob parseFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry) throws IOException
IOException
public PySparkJob.Builder newBuilderForType()
newBuilderForType
in interface Message
newBuilderForType
in interface MessageLite
public static PySparkJob.Builder newBuilder()
public static PySparkJob.Builder newBuilder(PySparkJob prototype)
public PySparkJob.Builder toBuilder()
toBuilder
in interface Message
toBuilder
in interface MessageLite
protected PySparkJob.Builder newBuilderForType(GeneratedMessageV3.BuilderParent parent)
newBuilderForType
in class GeneratedMessageV3
public static PySparkJob getDefaultInstance()
public static Parser<PySparkJob> parser()
public Parser<PySparkJob> getParserForType()
getParserForType
in interface Message
getParserForType
in interface MessageLite
getParserForType
in class GeneratedMessageV3
public PySparkJob getDefaultInstanceForType()
getDefaultInstanceForType
in interface MessageLiteOrBuilder
getDefaultInstanceForType
in interface MessageOrBuilder
Copyright © 2022 Google LLC. All rights reserved.