Skip to content

chore: Update discovery artifacts #2411

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
34 commits
Select commit Hold shift + click to select a range
9d6000f
feat(aiplatform): update the api
yoshi-automation Jun 4, 2024
494a29d
feat(analyticsadmin): update the api
yoshi-automation Jun 4, 2024
5afc401
feat(androidmanagement): update the api
yoshi-automation Jun 4, 2024
5bcc5d3
feat(backupdr): update the api
yoshi-automation Jun 4, 2024
32ddf52
feat(chromemanagement): update the api
yoshi-automation Jun 4, 2024
2b5c66b
feat(cloudbilling): update the api
yoshi-automation Jun 4, 2024
34314fb
feat(cloudfunctions): update the api
yoshi-automation Jun 4, 2024
d32e900
feat(cloudsearch): update the api
yoshi-automation Jun 4, 2024
4f7da21
feat(compute): update the api
yoshi-automation Jun 4, 2024
8087f14
feat(connectors): update the api
yoshi-automation Jun 4, 2024
4fb577d
feat(contactcenteraiplatform): update the api
yoshi-automation Jun 4, 2024
bb49784
feat(contactcenterinsights): update the api
yoshi-automation Jun 4, 2024
ac474a9
feat(datamigration): update the api
yoshi-automation Jun 4, 2024
d959b3d
feat(dataplex): update the api
yoshi-automation Jun 4, 2024
3abd0f4
feat(datastream): update the api
yoshi-automation Jun 4, 2024
2d79840
feat(dialogflow): update the api
yoshi-automation Jun 4, 2024
4522cd5
feat(discoveryengine): update the api
yoshi-automation Jun 4, 2024
a06827e
feat(documentai): update the api
yoshi-automation Jun 4, 2024
f7c50fd
feat(fcmdata): update the api
yoshi-automation Jun 4, 2024
0744228
feat(firebaseappcheck): update the api
yoshi-automation Jun 4, 2024
05c4657
feat(healthcare): update the api
yoshi-automation Jun 4, 2024
331029f
feat(iam): update the api
yoshi-automation Jun 4, 2024
8bd4954
feat(integrations): update the api
yoshi-automation Jun 4, 2024
b46b8b7
feat(migrationcenter): update the api
yoshi-automation Jun 4, 2024
ff49e0b
feat(networkconnectivity): update the api
yoshi-automation Jun 4, 2024
b56b2b1
feat(policyanalyzer): update the api
yoshi-automation Jun 4, 2024
a5e25b3
feat(resourcesettings): update the api
yoshi-automation Jun 4, 2024
81892c8
feat(run): update the api
yoshi-automation Jun 4, 2024
d0199ea
fix(secretmanager): update the api
yoshi-automation Jun 4, 2024
0cfcab3
feat(servicecontrol): update the api
yoshi-automation Jun 4, 2024
195cae3
feat(spanner): update the api
yoshi-automation Jun 4, 2024
9cef71c
feat(versionhistory): update the api
yoshi-automation Jun 4, 2024
6670b1e
feat(workflowexecutions): update the api
yoshi-automation Jun 4, 2024
26ed0d0
chore(docs): Add new discovery artifacts and artifacts with minor upd…
yoshi-automation Jun 4, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
Original file line number Diff line number Diff line change
Expand Up @@ -360,7 +360,7 @@ <h3>Method Details</h3>
},
],
&quot;healthProbe&quot;: { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # Exec specifies the action to take.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command.
&quot;command&quot;: [ # Command is the command line to execute inside the container, the working directory for the command is root (&#x27;/&#x27;) in the container&#x27;s filesystem. The command is simply exec&#x27;d, it is not run inside a shell, so traditional shell instructions (&#x27;|&#x27;, etc) won&#x27;t work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.
&quot;A String&quot;,
],
Expand All @@ -378,7 +378,7 @@ <h3>Method Details</h3>
&quot;predictRoute&quot;: &quot;A String&quot;, # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container&#x27;s IP address and port. Vertex AI then returns the container&#x27;s response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`&#x27;s ports field. If you don&#x27;t specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
&quot;sharedMemorySizeMb&quot;: &quot;A String&quot;, # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes.
&quot;startupProbe&quot;: { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # Exec specifies the action to take.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command.
&quot;command&quot;: [ # Command is the command line to execute inside the container, the working directory for the command is root (&#x27;/&#x27;) in the container&#x27;s filesystem. The command is simply exec&#x27;d, it is not run inside a shell, so traditional shell instructions (&#x27;|&#x27;, etc) won&#x27;t work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.
&quot;A String&quot;,
],
Expand Down Expand Up @@ -630,7 +630,7 @@ <h3>Method Details</h3>
},
],
&quot;healthProbe&quot;: { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # Exec specifies the action to take.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command.
&quot;command&quot;: [ # Command is the command line to execute inside the container, the working directory for the command is root (&#x27;/&#x27;) in the container&#x27;s filesystem. The command is simply exec&#x27;d, it is not run inside a shell, so traditional shell instructions (&#x27;|&#x27;, etc) won&#x27;t work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.
&quot;A String&quot;,
],
Expand All @@ -648,7 +648,7 @@ <h3>Method Details</h3>
&quot;predictRoute&quot;: &quot;A String&quot;, # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container&#x27;s IP address and port. Vertex AI then returns the container&#x27;s response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`&#x27;s ports field. If you don&#x27;t specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
&quot;sharedMemorySizeMb&quot;: &quot;A String&quot;, # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes.
&quot;startupProbe&quot;: { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # Exec specifies the action to take.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command.
&quot;command&quot;: [ # Command is the command line to execute inside the container, the working directory for the command is root (&#x27;/&#x27;) in the container&#x27;s filesystem. The command is simply exec&#x27;d, it is not run inside a shell, so traditional shell instructions (&#x27;|&#x27;, etc) won&#x27;t work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.
&quot;A String&quot;,
],
Expand Down Expand Up @@ -942,7 +942,7 @@ <h3>Method Details</h3>
},
],
&quot;healthProbe&quot;: { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # Exec specifies the action to take.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command.
&quot;command&quot;: [ # Command is the command line to execute inside the container, the working directory for the command is root (&#x27;/&#x27;) in the container&#x27;s filesystem. The command is simply exec&#x27;d, it is not run inside a shell, so traditional shell instructions (&#x27;|&#x27;, etc) won&#x27;t work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.
&quot;A String&quot;,
],
Expand All @@ -960,7 +960,7 @@ <h3>Method Details</h3>
&quot;predictRoute&quot;: &quot;A String&quot;, # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container&#x27;s IP address and port. Vertex AI then returns the container&#x27;s response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`&#x27;s ports field. If you don&#x27;t specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
&quot;sharedMemorySizeMb&quot;: &quot;A String&quot;, # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes.
&quot;startupProbe&quot;: { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # Exec specifies the action to take.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command.
&quot;command&quot;: [ # Command is the command line to execute inside the container, the working directory for the command is root (&#x27;/&#x27;) in the container&#x27;s filesystem. The command is simply exec&#x27;d, it is not run inside a shell, so traditional shell instructions (&#x27;|&#x27;, etc) won&#x27;t work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.
&quot;A String&quot;,
],
Expand Down Expand Up @@ -1225,7 +1225,7 @@ <h3>Method Details</h3>
},
],
&quot;healthProbe&quot;: { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # Exec specifies the action to take.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command.
&quot;command&quot;: [ # Command is the command line to execute inside the container, the working directory for the command is root (&#x27;/&#x27;) in the container&#x27;s filesystem. The command is simply exec&#x27;d, it is not run inside a shell, so traditional shell instructions (&#x27;|&#x27;, etc) won&#x27;t work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.
&quot;A String&quot;,
],
Expand All @@ -1243,7 +1243,7 @@ <h3>Method Details</h3>
&quot;predictRoute&quot;: &quot;A String&quot;, # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container&#x27;s IP address and port. Vertex AI then returns the container&#x27;s response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`&#x27;s ports field. If you don&#x27;t specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
&quot;sharedMemorySizeMb&quot;: &quot;A String&quot;, # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes.
&quot;startupProbe&quot;: { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # Exec specifies the action to take.
&quot;exec&quot;: { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command.
&quot;command&quot;: [ # Command is the command line to execute inside the container, the working directory for the command is root (&#x27;/&#x27;) in the container&#x27;s filesystem. The command is simply exec&#x27;d, it is not run inside a shell, so traditional shell instructions (&#x27;|&#x27;, etc) won&#x27;t work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.
&quot;A String&quot;,
],
Expand Down
Loading