- 0.58.0 (latest)
- 0.57.0
- 0.56.0
- 0.55.0
- 0.54.0
- 0.53.0
- 0.52.0
- 0.51.0
- 0.50.0
- 0.49.0
- 0.48.0
- 0.47.0
- 0.46.0
- 0.45.0
- 0.44.0
- 0.43.0
- 0.42.0
- 0.41.0
- 0.40.0
- 0.39.0
- 0.38.0
- 0.37.0
- 0.36.0
- 0.35.0
- 0.34.0
- 0.33.0
- 0.32.0
- 0.31.0
- 0.30.0
- 0.29.0
- 0.28.0
- 0.27.0
- 0.26.0
- 0.25.0
- 0.24.0
- 0.23.0
- 0.22.0
- 0.21.0
- 0.20.0
- 0.19.0
- 0.18.0
- 0.17.0
- 0.16.0
- 0.15.0
- 0.14.0
- 0.13.0
- 0.12.0
- 0.11.0
- 0.10.0
- 0.9.1
- 0.8.0
- 0.7.0
- 0.6.0
- 0.5.0
- 0.4.0
- 0.3.0
- 0.2.0
- 0.1.0
Reference documentation and code samples for the Vertex AI V1 API class Google::Cloud::AIPlatform::V1::DeployedModel.
A deployment of a Model. Endpoints contain one or more DeployedModels.
Inherits
- Object
Extended By
- Google::Protobuf::MessageExts::ClassMethods
Includes
- Google::Protobuf::MessageExts
Methods
#automatic_resources
def automatic_resources() -> ::Google::Cloud::AIPlatform::V1::AutomaticResources
- (::Google::Cloud::AIPlatform::V1::AutomaticResources) — A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
#automatic_resources=
def automatic_resources=(value) -> ::Google::Cloud::AIPlatform::V1::AutomaticResources
- value (::Google::Cloud::AIPlatform::V1::AutomaticResources) — A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
- (::Google::Cloud::AIPlatform::V1::AutomaticResources) — A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
#create_time
def create_time() -> ::Google::Protobuf::Timestamp
- (::Google::Protobuf::Timestamp) — Output only. Timestamp when the DeployedModel was created.
#dedicated_resources
def dedicated_resources() -> ::Google::Cloud::AIPlatform::V1::DedicatedResources
- (::Google::Cloud::AIPlatform::V1::DedicatedResources) — A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
#dedicated_resources=
def dedicated_resources=(value) -> ::Google::Cloud::AIPlatform::V1::DedicatedResources
- value (::Google::Cloud::AIPlatform::V1::DedicatedResources) — A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
- (::Google::Cloud::AIPlatform::V1::DedicatedResources) — A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
#disable_container_logging
def disable_container_logging() -> ::Boolean
-
(::Boolean) — For custom-trained Models and AutoML Tabular Models, the container of the
DeployedModel instances will send
stderr
andstdout
streams to Stackdriver Logging by default. Please note that the logs incur cost, which are subject to Cloud Logging pricing.User can disable container logging by setting this flag to true.
#disable_container_logging=
def disable_container_logging=(value) -> ::Boolean
-
value (::Boolean) — For custom-trained Models and AutoML Tabular Models, the container of the
DeployedModel instances will send
stderr
andstdout
streams to Stackdriver Logging by default. Please note that the logs incur cost, which are subject to Cloud Logging pricing.User can disable container logging by setting this flag to true.
-
(::Boolean) — For custom-trained Models and AutoML Tabular Models, the container of the
DeployedModel instances will send
stderr
andstdout
streams to Stackdriver Logging by default. Please note that the logs incur cost, which are subject to Cloud Logging pricing.User can disable container logging by setting this flag to true.
#display_name
def display_name() -> ::String
- (::String) — The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
#display_name=
def display_name=(value) -> ::String
- value (::String) — The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
- (::String) — The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
#enable_access_logging
def enable_access_logging() -> ::Boolean
-
(::Boolean) — These logs are like standard server access logs, containing
information like timestamp and latency for each prediction request.
Note that Stackdriver logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
#enable_access_logging=
def enable_access_logging=(value) -> ::Boolean
-
value (::Boolean) — These logs are like standard server access logs, containing
information like timestamp and latency for each prediction request.
Note that Stackdriver logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
-
(::Boolean) — These logs are like standard server access logs, containing
information like timestamp and latency for each prediction request.
Note that Stackdriver logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
#explanation_spec
def explanation_spec() -> ::Google::Cloud::AIPlatform::V1::ExplanationSpec
-
(::Google::Cloud::AIPlatform::V1::ExplanationSpec) — Explanation configuration for this DeployedModel.
When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.
#explanation_spec=
def explanation_spec=(value) -> ::Google::Cloud::AIPlatform::V1::ExplanationSpec
-
value (::Google::Cloud::AIPlatform::V1::ExplanationSpec) — Explanation configuration for this DeployedModel.
When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.
-
(::Google::Cloud::AIPlatform::V1::ExplanationSpec) — Explanation configuration for this DeployedModel.
When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.
#id
def id() -> ::String
-
(::String) — Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI
will generate a value for this ID.
This value should be 1-10 characters, and valid characters are /[0-9]/.
#id=
def id=(value) -> ::String
-
value (::String) — Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI
will generate a value for this ID.
This value should be 1-10 characters, and valid characters are /[0-9]/.
-
(::String) — Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI
will generate a value for this ID.
This value should be 1-10 characters, and valid characters are /[0-9]/.
#model
def model() -> ::String
- (::String) — Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint.
#model=
def model=(value) -> ::String
- value (::String) — Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint.
- (::String) — Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint.
#private_endpoints
def private_endpoints() -> ::Google::Cloud::AIPlatform::V1::PrivateEndpoints
- (::Google::Cloud::AIPlatform::V1::PrivateEndpoints) — Output only. Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if network is configured.
#service_account
def service_account() -> ::String
-
(::String) — The service account that the DeployedModel's container runs as. Specify the
email address of the service account. If this service account is not
specified, the container runs as a service account that doesn't have access
to the resource project.
Users deploying the Model must have the
iam.serviceAccounts.actAs
permission on this service account.
#service_account=
def service_account=(value) -> ::String
-
value (::String) — The service account that the DeployedModel's container runs as. Specify the
email address of the service account. If this service account is not
specified, the container runs as a service account that doesn't have access
to the resource project.
Users deploying the Model must have the
iam.serviceAccounts.actAs
permission on this service account.
-
(::String) — The service account that the DeployedModel's container runs as. Specify the
email address of the service account. If this service account is not
specified, the container runs as a service account that doesn't have access
to the resource project.
Users deploying the Model must have the
iam.serviceAccounts.actAs
permission on this service account.