Class: Google::Cloud::Dataproc::V1beta2::SparkJob

Inherits:
Object
  • Object
show all
Defined in:
lib/google/cloud/dataproc/v1beta2/doc/google/cloud/dataproc/v1beta2/jobs.rb

Overview

A Dataproc job for running Apache Spark applications on YARN. The specification of the main method to call to drive the job. Specify either the jar file that contains the main class or the main class name. To pass both a main jar and a main class in that jar, add the jar to CommonJob.jar_file_uris, and then specify the main class name in main_class.

Instance Attribute Summary collapse

Instance Attribute Details

#archive_urisArray<String>

Returns Optional. HCFS URIs of archives to be extracted in the working directory of Spark drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.

Returns:

  • (Array<String>)

    Optional. HCFS URIs of archives to be extracted in the working directory of Spark drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.



149
# File 'lib/google/cloud/dataproc/v1beta2/doc/google/cloud/dataproc/v1beta2/jobs.rb', line 149

class SparkJob; end

#argsArray<String>

Returns Optional. The arguments to pass to the driver. Do not include arguments, such as --conf, that can be set as job properties, since a collision may occur that causes an incorrect job submission.

Returns:

  • (Array<String>)

    Optional. The arguments to pass to the driver. Do not include arguments, such as --conf, that can be set as job properties, since a collision may occur that causes an incorrect job submission.



149
# File 'lib/google/cloud/dataproc/v1beta2/doc/google/cloud/dataproc/v1beta2/jobs.rb', line 149

class SparkJob; end

#file_urisArray<String>

Returns Optional. HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.

Returns:

  • (Array<String>)

    Optional. HCFS URIs of files to be copied to the working directory of Spark drivers and distributed tasks. Useful for naively parallel tasks.



149
# File 'lib/google/cloud/dataproc/v1beta2/doc/google/cloud/dataproc/v1beta2/jobs.rb', line 149

class SparkJob; end

#jar_file_urisArray<String>

Returns Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.

Returns:

  • (Array<String>)

    Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Spark driver and tasks.



149
# File 'lib/google/cloud/dataproc/v1beta2/doc/google/cloud/dataproc/v1beta2/jobs.rb', line 149

class SparkJob; end

#logging_configGoogle::Cloud::Dataproc::V1beta2::LoggingConfig

Returns Optional. The runtime log config for job execution.

Returns:



149
# File 'lib/google/cloud/dataproc/v1beta2/doc/google/cloud/dataproc/v1beta2/jobs.rb', line 149

class SparkJob; end

#main_classString

Returns The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in jar_file_uris.

Returns:

  • (String)

    The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in jar_file_uris.



149
# File 'lib/google/cloud/dataproc/v1beta2/doc/google/cloud/dataproc/v1beta2/jobs.rb', line 149

class SparkJob; end

#main_jar_file_uriString

Returns The HCFS URI of the jar file that contains the main class.

Returns:

  • (String)

    The HCFS URI of the jar file that contains the main class.



149
# File 'lib/google/cloud/dataproc/v1beta2/doc/google/cloud/dataproc/v1beta2/jobs.rb', line 149

class SparkJob; end

#propertiesHash{String => String}

Returns Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.

Returns:

  • (Hash{String => String})

    Optional. A mapping of property names to values, used to configure Spark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.



149
# File 'lib/google/cloud/dataproc/v1beta2/doc/google/cloud/dataproc/v1beta2/jobs.rb', line 149

class SparkJob; end