gcloud dataplex assets update - update a Dataplex asset resource
gcloud dataplex assets update (ASSET : --lake=LAKE --location=LOCATION --zone=ZONE) [--async] [--description=DESCRIPTION] [--display-name=DISPLAY_NAME] [--labels=[KEY=VALUE,...]] [--validate-only] [--[no-]discovery-enabled --discovery-exclude-patterns=[EXCLUDE_PATTERNS,...] --discovery-include-patterns=[INCLUDE_PATTERNS,...] --discovery-schedule=DISCOVERY_SCHEDULE --csv-delimiter=CSV_DELIMITER --[no-]csv-disable-type-inference --csv-encoding=CSV_ENCODING --csv-header-rows=CSV_HEADER_ROWS --[no-]json-disable-type-inference --json-encoding=JSON_ENCODING] [GCLOUD_WIDE_FLAG ...]
Update a Dataplex asset resource.
To update a Dataplex asset test-asset in zone test-zone in lake test-lake in location us-central1 to have the display name first-dataplex-asset and discovery include patterns abc, def, run:
$ gcloud dataplex assets update test-asset --location=us-central1 \ --lake=test-lake --zone=test-zone \ --display-name="first-dataplex-asset" \ --discovery-include-patterns=abc,def
- Assets resource - Arguments and flags that define the Dataplex asset you want
to update. The arguments in this group can be used to specify the attributes of this resource. (NOTE) Some attributes are not given arguments in this group but can be set in other ways. To set the project attribute:
- —
provide the argument asset on the command line with a fully specified name;
- —
provide the argument --project on the command line;
- —
set the property core/project.
This must be specified.
- ASSET
ID of the assets or fully qualified identifier for the assets. To set the asset attribute:
provide the argument asset on the command line.
This positional argument must be specified if any of the other arguments in this group are specified.
- --lake=LAKE
The identifier of the Dataplex lake resource. To set the lake attribute:
provide the argument asset on the command line with a fully specified name;
provide the argument --lake on the command line.
- --location=LOCATION
The location of the Dataplex resource. To set the location attribute:
provide the argument asset on the command line with a fully specified name;
provide the argument --location on the command line;
set the property dataplex/location.
- --zone=ZONE
The identifier of the Dataplex zone resource. To set the zone attribute:
provide the argument asset on the command line with a fully specified name;
provide the argument --zone on the command line.
- --async
Return immediately, without waiting for the operation in progress to complete.
- --description=DESCRIPTION
Description of the asset
- --display-name=DISPLAY_NAME
Display Name
- --labels=[KEY=VALUE,...]
List of label KEY=VALUE pairs to add.
Keys must start with a lowercase character and contain only hyphens (-), underscores (_), lowercase characters, and numbers. Values must contain only hyphens (-), underscores (_), lowercase characters, and numbers.
- --validate-only
Validate the update action, but don't actually perform it.
- Settings to manage the metadata discovery and publishing.
- --[no-]discovery-enabled
Whether discovery is enabled. Use --discovery-enabled to enable and --no-discovery-enabled to disable.
- --discovery-exclude-patterns=[EXCLUDE_PATTERNS,...]
The list of patterns to apply for selecting data to exclude during discovery. For Cloud Storage bucket assets, these are interpreted as glob patterns used to match object names. For BigQuery dataset assets, these are interpreted as patterns to match table names.
- --discovery-include-patterns=[INCLUDE_PATTERNS,...]
The list of patterns to apply for selecting data to include during discovery if only a subset of the data should considered. For Cloud Storage bucket assets, these are interpreted as glob patterns used to match object names. For BigQuery dataset assets, these are interpreted as patterns to match table names.
- Determines when discovery jobs are triggered.
- --discovery-schedule=DISCOVERY_SCHEDULE
Cron schedule https://en.wikipedia.org/wiki/Cron for running discovery jobs periodically. Discovery jobs must be scheduled at least 30 minutes apart.
- Describe data formats.
- Describe CSV and similar semi-structured data formats.
- --csv-delimiter=CSV_DELIMITER
The delimiter being used to separate values. This defaults to ','.
- --[no-]csv-disable-type-inference
Whether to disable the inference of data type for CSV data. If true, all columns will be registered as strings. Use --csv-disable-type-inference to enable and --no-csv-disable-type-inference to disable.
- --csv-encoding=CSV_ENCODING
The character encoding of the data. The default is UTF-8.
- --csv-header-rows=CSV_HEADER_ROWS
The number of rows to interpret as header rows that should be skipped when reading data rows.
- Describe JSON data format.
- --[no-]json-disable-type-inference
Whether to disable the inference of data type for Json data. If true, all columns will be registered as their primitive types (strings, number or boolean). Use --json-disable-type-inference to enable and --no-json-disable-type-inference to disable.
- --json-encoding=JSON_ENCODING
The character encoding of the data. The default is UTF-8.
These flags are available to all commands: --access-token-file, --account, --billing-project, --configuration, --flags-file, --flatten, --format, --help, --impersonate-service-account, --log-http, --project, --quiet, --trace-token, --user-output-enabled, --verbosity.
Run $ gcloud help for details.
This variant is also available:
$ gcloud alpha dataplex assets update