Class: Google::Cloud::AIPlatform::V1::GenerateContentResponse::PromptFeedback
- Inherits:
-
Object
- Object
- Google::Cloud::AIPlatform::V1::GenerateContentResponse::PromptFeedback
- Extended by:
- Protobuf::MessageExts::ClassMethods
- Includes:
- Protobuf::MessageExts
- Defined in:
- proto_docs/google/cloud/aiplatform/v1/prediction_service.rb
Overview
Content filter results for a prompt sent in the request.
Defined Under Namespace
Modules: BlockedReason
Instance Attribute Summary collapse
-
#block_reason ⇒ ::Google::Cloud::AIPlatform::V1::GenerateContentResponse::PromptFeedback::BlockedReason
readonly
Output only.
-
#block_reason_message ⇒ ::String
readonly
Output only.
-
#safety_ratings ⇒ ::Array<::Google::Cloud::AIPlatform::V1::SafetyRating>
readonly
Output only.
Instance Attribute Details
#block_reason ⇒ ::Google::Cloud::AIPlatform::V1::GenerateContentResponse::PromptFeedback::BlockedReason (readonly)
Returns Output only. Blocked reason.
607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 |
# File 'proto_docs/google/cloud/aiplatform/v1/prediction_service.rb', line 607 class PromptFeedback include ::Google::Protobuf::MessageExts extend ::Google::Protobuf::MessageExts::ClassMethods # Blocked reason enumeration. module BlockedReason # Unspecified blocked reason. BLOCKED_REASON_UNSPECIFIED = 0 # Candidates blocked due to safety. SAFETY = 1 # Candidates blocked due to other reason. OTHER = 2 # Candidates blocked due to the terms which are included from the # terminology blocklist. BLOCKLIST = 3 # Candidates blocked due to prohibited content. PROHIBITED_CONTENT = 4 # The user prompt was blocked by Model Armor. MODEL_ARMOR = 5 # The user prompt was blocked due to jailbreak. JAILBREAK = 6 end end |
#block_reason_message ⇒ ::String (readonly)
Returns Output only. A readable block reason message.
607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 |
# File 'proto_docs/google/cloud/aiplatform/v1/prediction_service.rb', line 607 class PromptFeedback include ::Google::Protobuf::MessageExts extend ::Google::Protobuf::MessageExts::ClassMethods # Blocked reason enumeration. module BlockedReason # Unspecified blocked reason. BLOCKED_REASON_UNSPECIFIED = 0 # Candidates blocked due to safety. SAFETY = 1 # Candidates blocked due to other reason. OTHER = 2 # Candidates blocked due to the terms which are included from the # terminology blocklist. BLOCKLIST = 3 # Candidates blocked due to prohibited content. PROHIBITED_CONTENT = 4 # The user prompt was blocked by Model Armor. MODEL_ARMOR = 5 # The user prompt was blocked due to jailbreak. JAILBREAK = 6 end end |
#safety_ratings ⇒ ::Array<::Google::Cloud::AIPlatform::V1::SafetyRating> (readonly)
Returns Output only. Safety ratings.
607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 |
# File 'proto_docs/google/cloud/aiplatform/v1/prediction_service.rb', line 607 class PromptFeedback include ::Google::Protobuf::MessageExts extend ::Google::Protobuf::MessageExts::ClassMethods # Blocked reason enumeration. module BlockedReason # Unspecified blocked reason. BLOCKED_REASON_UNSPECIFIED = 0 # Candidates blocked due to safety. SAFETY = 1 # Candidates blocked due to other reason. OTHER = 2 # Candidates blocked due to the terms which are included from the # terminology blocklist. BLOCKLIST = 3 # Candidates blocked due to prohibited content. PROHIBITED_CONTENT = 4 # The user prompt was blocked by Model Armor. MODEL_ARMOR = 5 # The user prompt was blocked due to jailbreak. JAILBREAK = 6 end end |