Class: Google::Cloud::AIPlatform::V1::MachineSpec

Inherits:
Object
  • Object
show all
Extended by:
Protobuf::MessageExts::ClassMethods
Includes:
Protobuf::MessageExts
Defined in:
proto_docs/google/cloud/aiplatform/v1/machine_resources.rb

Overview

Specification of a single machine.

Instance Attribute Summary collapse

Instance Attribute Details

#accelerator_count::Integer

Returns The number of accelerators to attach to the machine.

Returns:

  • (::Integer)

    The number of accelerators to attach to the machine.



72
73
74
75
# File 'proto_docs/google/cloud/aiplatform/v1/machine_resources.rb', line 72

class MachineSpec
  include ::Google::Protobuf::MessageExts
  extend ::Google::Protobuf::MessageExts::ClassMethods
end

#accelerator_type::Google::Cloud::AIPlatform::V1::AcceleratorType

Returns Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count.

Returns:



72
73
74
75
# File 'proto_docs/google/cloud/aiplatform/v1/machine_resources.rb', line 72

class MachineSpec
  include ::Google::Protobuf::MessageExts
  extend ::Google::Protobuf::MessageExts::ClassMethods
end

#gpu_partition_size::String

Returns Optional. Immutable. The Nvidia GPU partition size.

When specified, the requested accelerators will be partitioned into smaller GPU partitions. For example, if the request is for 8 units of NVIDIA A100 GPUs, and gpu_partition_size="1g.10gb", the service will create 8 * 7 = 56 partitioned MIG instances.

The partition size must be a value supported by the requested accelerator. Refer to Nvidia GPU Partitioning for the available partition sizes.

If set, the accelerator_count should be set to 1.

Returns:

  • (::String)

    Optional. Immutable. The Nvidia GPU partition size.

    When specified, the requested accelerators will be partitioned into smaller GPU partitions. For example, if the request is for 8 units of NVIDIA A100 GPUs, and gpu_partition_size="1g.10gb", the service will create 8 * 7 = 56 partitioned MIG instances.

    The partition size must be a value supported by the requested accelerator. Refer to Nvidia GPU Partitioning for the available partition sizes.

    If set, the accelerator_count should be set to 1.



72
73
74
75
# File 'proto_docs/google/cloud/aiplatform/v1/machine_resources.rb', line 72

class MachineSpec
  include ::Google::Protobuf::MessageExts
  extend ::Google::Protobuf::MessageExts::ClassMethods
end

#machine_type::String

Returns Immutable. The type of the machine.

See the list of machine types supported for prediction

See the list of machine types supported for custom training.

For DeployedModel this field is optional, and the default value is n1-standard-2. For BatchPredictionJob or as part of WorkerPoolSpec this field is required.

Returns:



72
73
74
75
# File 'proto_docs/google/cloud/aiplatform/v1/machine_resources.rb', line 72

class MachineSpec
  include ::Google::Protobuf::MessageExts
  extend ::Google::Protobuf::MessageExts::ClassMethods
end

#reservation_affinity::Google::Cloud::AIPlatform::V1::ReservationAffinity

Returns Optional. Immutable. Configuration controlling how this resource pool consumes reservation.

Returns:



72
73
74
75
# File 'proto_docs/google/cloud/aiplatform/v1/machine_resources.rb', line 72

class MachineSpec
  include ::Google::Protobuf::MessageExts
  extend ::Google::Protobuf::MessageExts::ClassMethods
end

#tpu_topology::String

Returns Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").

Returns:

  • (::String)

    Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").



72
73
74
75
# File 'proto_docs/google/cloud/aiplatform/v1/machine_resources.rb', line 72

class MachineSpec
  include ::Google::Protobuf::MessageExts
  extend ::Google::Protobuf::MessageExts::ClassMethods
end