Class: Google::Apis::DataprocV1::Job
- Inherits:
-
Object
- Object
- Google::Apis::DataprocV1::Job
- Includes:
- Core::Hashable, Core::JsonObjectSupport
- Defined in:
- generated/google/apis/dataproc_v1/classes.rb,
generated/google/apis/dataproc_v1/representations.rb,
generated/google/apis/dataproc_v1/representations.rb
Overview
A Cloud Dataproc job resource.
Instance Attribute Summary collapse
-
#driver_control_files_uri ⇒ String
Output-only If present, the location of miscellaneous control files which may be used as part of job setup and handling.
-
#driver_output_resource_uri ⇒ String
Output-only A URI pointing to the location of the stdout of the job's driver program.
-
#hadoop_job ⇒ Google::Apis::DataprocV1::HadoopJob
A Cloud Dataproc job for running Apache Hadoop MapReduce (https://hadoop. apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/ MapReduceTutorial.html) jobs on Apache Hadoop YARN (https://hadoop.apache.org/ docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html).
-
#hive_job ⇒ Google::Apis::DataprocV1::HiveJob
A Cloud Dataproc job for running Apache Hive (https://hive.apache.org/) queries on YARN.
-
#labels ⇒ Hash<String,String>
Optional The labels to associate with this job.
-
#pig_job ⇒ Google::Apis::DataprocV1::PigJob
A Cloud Dataproc job for running Apache Pig (https://pig.apache.org/) queries on YARN.
-
#placement ⇒ Google::Apis::DataprocV1::JobPlacement
Cloud Dataproc job config.
-
#pyspark_job ⇒ Google::Apis::DataprocV1::PySparkJob
A Cloud Dataproc job for running Apache PySpark (https://spark.apache.org/docs/ 0.9.0/python-programming-guide.html) applications on YARN.
-
#reference ⇒ Google::Apis::DataprocV1::JobReference
Encapsulates the full scoping used to reference a job.
-
#scheduling ⇒ Google::Apis::DataprocV1::JobScheduling
Job scheduling options.Beta Feature: These options are available for testing purposes only.
-
#spark_job ⇒ Google::Apis::DataprocV1::SparkJob
A Cloud Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN.
-
#spark_sql_job ⇒ Google::Apis::DataprocV1::SparkSqlJob
A Cloud Dataproc job for running Apache Spark SQL (http://spark.apache.org/sql/ ) queries.
-
#status ⇒ Google::Apis::DataprocV1::JobStatus
Cloud Dataproc job status.
-
#status_history ⇒ Array<Google::Apis::DataprocV1::JobStatus>
Output-only The previous job status.
-
#yarn_applications ⇒ Array<Google::Apis::DataprocV1::YarnApplication>
Output-only The collection of YARN applications spun up by this job.Beta Feature: This report is available for testing purposes only.
Instance Method Summary collapse
-
#initialize(**args) ⇒ Job
constructor
A new instance of Job.
-
#update!(**args) ⇒ Object
Update properties of this object.
Methods included from Core::JsonObjectSupport
Methods included from Core::Hashable
Constructor Details
#initialize(**args) ⇒ Job
Returns a new instance of Job
404 405 406 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 404 def initialize(**args) update!(**args) end |
Instance Attribute Details
#driver_control_files_uri ⇒ String
Output-only If present, the location of miscellaneous control files which may
be used as part of job setup and handling. If not present, control files may
be placed in the same location as driver_output_uri.
Corresponds to the JSON property driverControlFilesUri
326 327 328 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 326 def driver_control_files_uri @driver_control_files_uri end |
#driver_output_resource_uri ⇒ String
Output-only A URI pointing to the location of the stdout of the job's driver
program.
Corresponds to the JSON property driverOutputResourceUri
359 360 361 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 359 def driver_output_resource_uri @driver_output_resource_uri end |
#hadoop_job ⇒ Google::Apis::DataprocV1::HadoopJob
A Cloud Dataproc job for running Apache Hadoop MapReduce (https://hadoop.
apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/
MapReduceTutorial.html) jobs on Apache Hadoop YARN (https://hadoop.apache.org/
docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html).
Corresponds to the JSON property hadoopJob
402 403 404 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 402 def hadoop_job @hadoop_job end |
#hive_job ⇒ Google::Apis::DataprocV1::HiveJob
A Cloud Dataproc job for running Apache Hive (https://hive.apache.org/)
queries on YARN.
Corresponds to the JSON property hiveJob
344 345 346 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 344 def hive_job @hive_job end |
#labels ⇒ Hash<String,String>
Optional The labels to associate with this job. Label keys must contain 1 to
63 characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.
txt). Label values may be empty, but, if present, must contain 1 to 63
characters, and must conform to RFC 1035 (https://www.ietf.org/rfc/rfc1035.txt)
. No more than 32 labels can be associated with a job.
Corresponds to the JSON property labels
353 354 355 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 353 def labels @labels end |
#pig_job ⇒ Google::Apis::DataprocV1::PigJob
A Cloud Dataproc job for running Apache Pig (https://pig.apache.org/) queries
on YARN.
Corresponds to the JSON property pigJob
338 339 340 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 338 def pig_job @pig_job end |
#placement ⇒ Google::Apis::DataprocV1::JobPlacement
Cloud Dataproc job config.
Corresponds to the JSON property placement
319 320 321 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 319 def placement @placement end |
#pyspark_job ⇒ Google::Apis::DataprocV1::PySparkJob
A Cloud Dataproc job for running Apache PySpark (https://spark.apache.org/docs/
0.9.0/python-programming-guide.html) applications on YARN.
Corresponds to the JSON property pysparkJob
389 390 391 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 389 def pyspark_job @pyspark_job end |
#reference ⇒ Google::Apis::DataprocV1::JobReference
Encapsulates the full scoping used to reference a job.
Corresponds to the JSON property reference
394 395 396 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 394 def reference @reference end |
#scheduling ⇒ Google::Apis::DataprocV1::JobScheduling
Job scheduling options.Beta Feature: These options are available for testing
purposes only. They may be changed before final release.
Corresponds to the JSON property scheduling
332 333 334 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 332 def scheduling @scheduling end |
#spark_job ⇒ Google::Apis::DataprocV1::SparkJob
A Cloud Dataproc job for running Apache Spark (http://spark.apache.org/)
applications on YARN.
Corresponds to the JSON property sparkJob
365 366 367 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 365 def spark_job @spark_job end |
#spark_sql_job ⇒ Google::Apis::DataprocV1::SparkSqlJob
A Cloud Dataproc job for running Apache Spark SQL (http://spark.apache.org/sql/
) queries.
Corresponds to the JSON property sparkSqlJob
371 372 373 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 371 def spark_sql_job @spark_sql_job end |
#status ⇒ Google::Apis::DataprocV1::JobStatus
Cloud Dataproc job status.
Corresponds to the JSON property status
314 315 316 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 314 def status @status end |
#status_history ⇒ Array<Google::Apis::DataprocV1::JobStatus>
Output-only The previous job status.
Corresponds to the JSON property statusHistory
376 377 378 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 376 def status_history @status_history end |
#yarn_applications ⇒ Array<Google::Apis::DataprocV1::YarnApplication>
Output-only The collection of YARN applications spun up by this job.Beta
Feature: This report is available for testing purposes only. It may be changed
before final release.
Corresponds to the JSON property yarnApplications
383 384 385 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 383 def yarn_applications @yarn_applications end |
Instance Method Details
#update!(**args) ⇒ Object
Update properties of this object
409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 |
# File 'generated/google/apis/dataproc_v1/classes.rb', line 409 def update!(**args) @status = args[:status] if args.key?(:status) @placement = args[:placement] if args.key?(:placement) @driver_control_files_uri = args[:driver_control_files_uri] if args.key?(:driver_control_files_uri) @scheduling = args[:scheduling] if args.key?(:scheduling) @pig_job = args[:pig_job] if args.key?(:pig_job) @hive_job = args[:hive_job] if args.key?(:hive_job) @labels = args[:labels] if args.key?(:labels) @driver_output_resource_uri = args[:driver_output_resource_uri] if args.key?(:driver_output_resource_uri) @spark_job = args[:spark_job] if args.key?(:spark_job) @spark_sql_job = args[:spark_sql_job] if args.key?(:spark_sql_job) @status_history = args[:status_history] if args.key?(:status_history) @yarn_applications = args[:yarn_applications] if args.key?(:yarn_applications) @pyspark_job = args[:pyspark_job] if args.key?(:pyspark_job) @reference = args[:reference] if args.key?(:reference) @hadoop_job = args[:hadoop_job] if args.key?(:hadoop_job) end |