Skip to content

Commit 987da0b

Browse files
feat(dlp): update the api
#### dlp:v2 The following keys were added: - schemas.GooglePrivacyDlpV2Result.properties.numRowsProcessed (Total Keys: 2)
1 parent 8472bf0 commit 987da0b

10 files changed

+36
-22
lines changed

docs/dyn/dlp_v2.organizations.locations.discoveryConfigs.html

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -116,7 +116,7 @@ <h3>Method Details</h3>
116116
&quot;actions&quot;: [ # Actions to execute at the completion of scanning.
117117
{ # A task to execute when a data profile has been generated.
118118
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
119-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
119+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
120120
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
121121
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
122122
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.
@@ -278,7 +278,7 @@ <h3>Method Details</h3>
278278
&quot;actions&quot;: [ # Actions to execute at the completion of scanning.
279279
{ # A task to execute when a data profile has been generated.
280280
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
281-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
281+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
282282
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
283283
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
284284
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.
@@ -464,7 +464,7 @@ <h3>Method Details</h3>
464464
&quot;actions&quot;: [ # Actions to execute at the completion of scanning.
465465
{ # A task to execute when a data profile has been generated.
466466
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
467-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
467+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
468468
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
469469
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
470470
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.
@@ -637,7 +637,7 @@ <h3>Method Details</h3>
637637
&quot;actions&quot;: [ # Actions to execute at the completion of scanning.
638638
{ # A task to execute when a data profile has been generated.
639639
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
640-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
640+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
641641
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
642642
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
643643
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.
@@ -818,7 +818,7 @@ <h3>Method Details</h3>
818818
&quot;actions&quot;: [ # Actions to execute at the completion of scanning.
819819
{ # A task to execute when a data profile has been generated.
820820
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
821-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
821+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
822822
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
823823
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
824824
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.
@@ -981,7 +981,7 @@ <h3>Method Details</h3>
981981
&quot;actions&quot;: [ # Actions to execute at the completion of scanning.
982982
{ # A task to execute when a data profile has been generated.
983983
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
984-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
984+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
985985
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
986986
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
987987
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.

docs/dyn/dlp_v2.organizations.locations.dlpJobs.html

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -3244,6 +3244,7 @@ <h3>Method Details</h3>
32443244
},
32453245
},
32463246
],
3247+
&quot;numRowsProcessed&quot;: &quot;A String&quot;, # Number of rows scanned post sampling and time filtering (Applicable for row based stores such as BigQuery).
32473248
&quot;processedBytes&quot;: &quot;A String&quot;, # Total size in bytes that were processed.
32483249
&quot;totalEstimatedBytes&quot;: &quot;A String&quot;, # Estimate of the number of bytes to process.
32493250
},

docs/dyn/dlp_v2.organizations.locations.tableDataProfiles.html

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -133,7 +133,7 @@ <h3>Method Details</h3>
133133
&quot;dataProfileActions&quot;: [ # Actions to execute at the completion of the job.
134134
{ # A task to execute when a data profile has been generated.
135135
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
136-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
136+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
137137
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
138138
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
139139
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.
@@ -170,7 +170,7 @@ <h3>Method Details</h3>
170170
&quot;actions&quot;: [ # Actions to execute at the completion of scanning.
171171
{ # A task to execute when a data profile has been generated.
172172
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
173-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
173+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
174174
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
175175
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
176176
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.
@@ -596,7 +596,7 @@ <h3>Method Details</h3>
596596
&quot;dataProfileActions&quot;: [ # Actions to execute at the completion of the job.
597597
{ # A task to execute when a data profile has been generated.
598598
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
599-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
599+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
600600
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
601601
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
602602
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.
@@ -633,7 +633,7 @@ <h3>Method Details</h3>
633633
&quot;actions&quot;: [ # Actions to execute at the completion of scanning.
634634
{ # A task to execute when a data profile has been generated.
635635
&quot;exportData&quot;: { # If set, the detailed data profiles will be persisted to the location of your choice whenever updated. # Export data profiles into a provided location.
636-
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in a new row in BigQuery.
636+
&quot;profileTable&quot;: { # Message defining the location of a BigQuery table. A table is uniquely identified by its project_id, dataset_id, and table_name. Within a query a table is often referenced with a string in the format of: `:.` or `..`. # Store all table and column profiles in an existing table or a new table in an existing dataset. Each re-generation will result in new rows in BigQuery. Data is inserted using [streaming insert](https://cloud.google.com/blog/products/bigquery/life-of-a-bigquery-streaming-insert) and so data may be in the buffer for a period of time after the profile has finished. The Pub/Sub notification is sent before the streaming buffer is guaranteed to be written, so data may not be instantly visible to queries by the time your topic receives the Pub/Sub notification.
637637
&quot;datasetId&quot;: &quot;A String&quot;, # Dataset ID of the table.
638638
&quot;projectId&quot;: &quot;A String&quot;, # The Google Cloud Platform project ID of the project containing the table. If omitted, project ID is inferred from the API call.
639639
&quot;tableId&quot;: &quot;A String&quot;, # Name of the table.

0 commit comments

Comments
 (0)