Hi there! Are you looking for the official Deno documentation? Try docs.deno.com for all your Deno learning needs.

GoogleCloudAiplatformV1DeployedModel

import type { GoogleCloudAiplatformV1DeployedModel } from "https://googleapis.deno.dev/v1/aiplatform:v1.ts";

A deployment of a Model. Endpoints contain one or more DeployedModels.

interface GoogleCloudAiplatformV1DeployedModel {
readonly createTime?: Date;
disableContainerLogging?: boolean;
disableExplanations?: boolean;
displayName?: string;
enableAccessLogging?: boolean;
id?: string;
model?: string;
readonly modelVersionId?: string;
readonly privateEndpoints?: GoogleCloudAiplatformV1PrivateEndpoints;
serviceAccount?: string;
sharedResources?: string;
}

§Properties

§

A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.

§
readonly createTime?: Date
[src]

Output only. Timestamp when the DeployedModel was created.

§

A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.

§
disableContainerLogging?: boolean
[src]

For custom-trained Models and AutoML Tabular Models, the container of the DeployedModel instances will send stderr and stdout streams to Cloud Logging by default. Please note that the logs incur cost, which are subject to Cloud Logging pricing. User can disable container logging by setting this flag to true.

§
disableExplanations?: boolean
[src]

If true, deploy the model without explainable feature, regardless the existence of Model.explanation_spec or explanation_spec.

§
displayName?: string
[src]

The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.

§
enableAccessLogging?: boolean
[src]

If true, online prediction access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each prediction request. Note that logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.

§

Explanation configuration for this DeployedModel. When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration.

§
id?: string
[src]

Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are /[0-9]/.

§
model?: string
[src]

Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: projects/{project}/locations/{location}/models/{model}@2 or projects/{project}/locations/{location}/models/{model}@golden if no version is specified, the default version will be deployed.

§
readonly modelVersionId?: string
[src]

Output only. The version ID of the model that is deployed.

§

Output only. Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if network is configured.

§
serviceAccount?: string
[src]

The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have the iam.serviceAccounts.actAs permission on this service account.

§
sharedResources?: string
[src]

The resource name of the shared DeploymentResourcePool to deploy on. Format: projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}