update all json files

This commit is contained in:
Sebastian Thiel
2020-04-12 18:55:37 +08:00
parent ea3b428364
commit aacc30f08d
260 changed files with 138411 additions and 26293 deletions

View File

@@ -21,6 +21,7 @@
},
"id": "storagetransfer:v1",
"kind": "discovery#restDescription",
"mtlsRootUrl": "https://storagetransfer.mtls.googleapis.com/",
"name": "storagetransfer",
"ownerDomain": "google.com",
"ownerName": "Google",
@@ -163,7 +164,7 @@
"jobName": {
"description": "Required. The job to get.",
"location": "path",
"pattern": "^transferJobs/.+$",
"pattern": "^transferJobs/.*$",
"required": true,
"type": "string"
},
@@ -189,7 +190,7 @@
"parameterOrder": [],
"parameters": {
"filter": {
"description": "Required. A list of query parameters specified as JSON text in the form of:\n{\"project_id\":\"my_project_id\",\n \"job_names\":[\"jobid1\",\"jobid2\",...],\n \"job_statuses\":[\"status1\",\"status2\",...]}.\nSince `job_names` and `job_statuses` support multiple values, their values\nmust be specified with array notation. `project_id` is required.\n`job_names` and `job_statuses` are optional. The valid values for\n`job_statuses` are case-insensitive: `ENABLED`, `DISABLED`, and `DELETED`.",
"description": "Required. A list of query parameters specified as JSON text in the form of:\n{\"project<span>_</span>id\":\"my_project_id\",\n \"job_names\":[\"jobid1\",\"jobid2\",...],\n \"job_statuses\":[\"status1\",\"status2\",...]}.\nSince `job_names` and `job_statuses` support multiple values, their values\nmust be specified with array notation. `project`<span>`_`</span>`id` is\nrequired. `job_names` and `job_statuses` are optional. The valid values\nfor `job_statuses` are case-insensitive:\nENABLED,\nDISABLED, and\nDELETED.",
"location": "query",
"type": "string"
},
@@ -214,7 +215,7 @@
]
},
"patch": {
"description": "Updates a transfer job. Updating a job's transfer spec does not affect\ntransfer operations that are running already. Updating the scheduling\nof a job is not allowed.",
"description": "Updates a transfer job. Updating a job's transfer spec does not affect\ntransfer operations that are running already. Updating a job's schedule\nis not allowed.\n\n**Note:** The job's status field can be modified\nusing this RPC (for example, to set a job's status to\nDELETED,\nDISABLED, or\nENABLED).",
"flatPath": "v1/transferJobs/{transferJobsId}",
"httpMethod": "PATCH",
"id": "storagetransfer.transferJobs.patch",
@@ -225,7 +226,7 @@
"jobName": {
"description": "Required. The name of job to update.",
"location": "path",
"pattern": "^transferJobs/.+$",
"pattern": "^transferJobs/.*$",
"required": true,
"type": "string"
}
@@ -257,7 +258,7 @@
"name": {
"description": "The name of the operation resource to be cancelled.",
"location": "path",
"pattern": "^transferOperations/.+$",
"pattern": "^transferOperations/.*$",
"required": true,
"type": "string"
}
@@ -270,31 +271,6 @@
"https://www.googleapis.com/auth/cloud-platform"
]
},
"delete": {
"description": "This method is not supported and the server returns `UNIMPLEMENTED`.",
"flatPath": "v1/transferOperations/{transferOperationsId}",
"httpMethod": "DELETE",
"id": "storagetransfer.transferOperations.delete",
"parameterOrder": [
"name"
],
"parameters": {
"name": {
"description": "The name of the operation resource to be deleted.",
"location": "path",
"pattern": "^transferOperations/.+$",
"required": true,
"type": "string"
}
},
"path": "v1/{+name}",
"response": {
"$ref": "Empty"
},
"scopes": [
"https://www.googleapis.com/auth/cloud-platform"
]
},
"get": {
"description": "Gets the latest state of a long-running operation. Clients can use this\nmethod to poll the operation result at intervals as recommended by the API\nservice.",
"flatPath": "v1/transferOperations/{transferOperationsId}",
@@ -307,7 +283,7 @@
"name": {
"description": "The name of the operation resource.",
"location": "path",
"pattern": "^transferOperations/.+$",
"pattern": "^transferOperations/.*$",
"required": true,
"type": "string"
}
@@ -321,7 +297,7 @@
]
},
"list": {
"description": "Lists operations that match the specified filter in the request. If the\nserver doesn't support this method, it returns `UNIMPLEMENTED`.\n\nNOTE: the `name` binding allows API services to override the binding\nto use different resource name schemes, such as `users/*/operations`. To\noverride the binding, API services can add a binding such as\n`\"/v1/{name=users/*}/operations\"` to their service configuration.\nFor backwards compatibility, the default name includes the operations\ncollection id, however overriding users must ensure the name binding\nis the parent resource, without the operations collection id.",
"description": "Lists transfer operations.",
"flatPath": "v1/transferOperations",
"httpMethod": "GET",
"id": "storagetransfer.transferOperations.list",
@@ -330,7 +306,7 @@
],
"parameters": {
"filter": {
"description": "Required. A list of query parameters specified as JSON text in the form of: {\"project_id\":\"my_project_id\",\n \"job_names\":[\"jobid1\",\"jobid2\",...],\n \"operation_names\":[\"opid1\",\"opid2\",...],\n \"transfer_statuses\":[\"status1\",\"status2\",...]}.\nSince `job_names`, `operation_names`, and `transfer_statuses` support multiple values, they must be specified with array notation. `project_id` is required. `job_names`, `operation_names`, and `transfer_statuses` are optional. The valid values for `transfer_statuses` are case-insensitive: `IN_PROGRESS`, `PAUSED`, `SUCCESS`, `FAILED`, and `ABORTED`.",
"description": "Required. A list of query parameters specified as JSON text in the form of: {\"project<span>_</span>id\":\"my_project_id\",\n \"job_names\":[\"jobid1\",\"jobid2\",...],\n \"operation_names\":[\"opid1\",\"opid2\",...],\n \"transfer_statuses\":[\"status1\",\"status2\",...]}.\nSince `job_names`, `operation_names`, and `transfer_statuses` support multiple values, they must be specified with array notation. `project`<span>`_`</span>`id` is required. `job_names`, `operation_names`, and `transfer_statuses` are optional. The valid values for `transfer_statuses` are case-insensitive: IN_PROGRESS, PAUSED, SUCCESS, FAILED, and ABORTED.",
"location": "query",
"type": "string"
},
@@ -373,7 +349,7 @@
"name": {
"description": "Required. The name of the transfer operation.",
"location": "path",
"pattern": "^transferOperations/.+$",
"pattern": "^transferOperations/.*$",
"required": true,
"type": "string"
}
@@ -401,7 +377,7 @@
"name": {
"description": "Required. The name of the transfer operation.",
"location": "path",
"pattern": "^transferOperations/.+$",
"pattern": "^transferOperations/.*$",
"required": true,
"type": "string"
}
@@ -420,11 +396,11 @@
}
}
},
"revision": "20190702",
"revision": "20200405",
"rootUrl": "https://storagetransfer.googleapis.com/",
"schemas": {
"AwsAccessKey": {
"description": "AWS access key (see\n[AWS Security\nCredentials](http://docs.aws.amazon.com/general/latest/gr/aws-security-credentials.html)).",
"description": "AWS access key (see\n[AWS Security\nCredentials](https://docs.aws.amazon.com/general/latest/gr/aws-security-credentials.html)).",
"id": "AwsAccessKey",
"properties": {
"accessKeyId": {
@@ -447,7 +423,37 @@
"description": "Required. AWS access key used to sign the API requests to the AWS S3\nbucket. Permissions on the bucket must be granted to the access ID of the\nAWS access key."
},
"bucketName": {
"description": "Required. S3 Bucket name (see\n[Creating a\nbucket](http://docs.aws.amazon.com/AmazonS3/latest/dev/create-bucket-get-location-example.html)).",
"description": "Required. S3 Bucket name (see\n[Creating a\nbucket](https://docs.aws.amazon.com/AmazonS3/latest/dev/create-bucket-get-location-example.html)).",
"type": "string"
}
},
"type": "object"
},
"AzureBlobStorageData": {
"description": "An AzureBlobStorageData resource can be a data source, but not a data sink.\nAn AzureBlobStorageData resource represents one Azure container. The storage\naccount determines the [Azure\nendpoint](https://docs.microsoft.com/en-us/azure/storage/common/storage-create-storage-account#storage-account-endpoints).\nIn an AzureBlobStorageData resource, a blobs's name is the [Azure Blob\nStorage blob's key\nname](https://docs.microsoft.com/en-us/rest/api/storageservices/naming-and-referencing-containers--blobs--and-metadata#blob-names).",
"id": "AzureBlobStorageData",
"properties": {
"azureCredentials": {
"$ref": "AzureCredentials",
"description": "Required. Credentials used to authenticate API requests to Azure."
},
"container": {
"description": "Required. The container to transfer from the Azure Storage account.",
"type": "string"
},
"storageAccount": {
"description": "Required. The name of the Azure Storage account.",
"type": "string"
}
},
"type": "object"
},
"AzureCredentials": {
"description": "Azure credentials",
"id": "AzureCredentials",
"properties": {
"sasToken": {
"description": "Required. Azure shared access signature. (see\n[Grant limited access to Azure Storage resources using shared access\nsignatures\n(SAS)](https://docs.microsoft.com/en-us/azure/storage/common/storage-sas-overview)).",
"type": "string"
}
},
@@ -551,7 +557,7 @@
"type": "string"
},
"errorLogEntries": {
"description": "Error samples.\n\nNo more than 100 error log entries may be recorded for a given\nerror code for a single task.",
"description": "Error samples.\n\nAt most 5 error log entries will be recorded for a given\nerror code for a single transfer operation.",
"items": {
"$ref": "ErrorLogEntry"
},
@@ -561,11 +567,11 @@
"type": "object"
},
"GcsData": {
"description": "In a GcsData resource, an object's name is the Google Cloud Storage object's\nname and its `lastModificationTime` refers to the object's updated time,\nwhich changes when the content or the metadata of the object is updated.",
"description": "In a GcsData resource, an object's name is the Cloud Storage object's\nname and its \"last modification time\" refers to the object's `updated`\nproperty of Cloud Storage objects, which changes when the content or the\nmetadata of the object is updated.",
"id": "GcsData",
"properties": {
"bucketName": {
"description": "Required. Google Cloud Storage bucket name (see\n[Bucket Name\nRequirements](https://cloud.google.com/storage/docs/naming#requirements)).",
"description": "Required. Cloud Storage bucket name (see\n[Bucket Name\nRequirements](https://cloud.google.com/storage/docs/naming#requirements)).",
"type": "string"
}
},
@@ -583,7 +589,7 @@
"type": "object"
},
"HttpData": {
"description": "An HttpData resource specifies a list of objects on the web to be transferred\nover HTTP. The information of the objects to be transferred is contained in\na file referenced by a URL. The first line in the file must be\n\"TsvHttpData-1.0\", which specifies the format of the file. Subsequent lines\nspecify the information of the list of objects, one object per list entry.\nEach entry has the following tab-delimited fields:\n\n* HTTP URL - The location of the object.\n\n* Length - The size of the object in bytes.\n\n* MD5 - The base64-encoded MD5 hash of the object.\n\nFor an example of a valid TSV file, see\n[Transferring data from\nURLs](https://cloud.google.com/storage/transfer/create-url-list).\n\nWhen transferring data based on a URL list, keep the following in mind:\n\n* When an object located at `http(s)://hostname:port/<URL-path>` is\ntransferred to a data sink, the name of the object at the data sink is\n`<hostname>/<URL-path>`.\n\n* If the specified size of an object does not match the actual size of the\nobject fetched, the object will not be transferred.\n\n* If the specified MD5 does not match the MD5 computed from the transferred\nbytes, the object transfer will fail. For more information, see\n[Generating MD5 hashes](https://cloud.google.com/storage/transfer/#md5)\n\n* Ensure that each URL you specify is publicly accessible. For\nexample, in Google Cloud Storage you can\n[share an object publicly]\n(https://cloud.google.com/storage/docs/cloud-console#_sharingdata) and get\na link to it.\n\n* Storage Transfer Service obeys `robots.txt` rules and requires the source\nHTTP server to support `Range` requests and to return a `Content-Length`\nheader in each response.\n\n* [ObjectConditions](#ObjectConditions) have no effect when filtering objects\nto transfer.",
"description": "An HttpData resource specifies a list of objects on the web to be transferred\nover HTTP. The information of the objects to be transferred is contained in\na file referenced by a URL. The first line in the file must be\n`\"TsvHttpData-1.0\"`, which specifies the format of the file. Subsequent\nlines specify the information of the list of objects, one object per list\nentry. Each entry has the following tab-delimited fields:\n\n* **HTTP URL** \u2014 The location of the object.\n\n* **Length** \u2014 The size of the object in bytes.\n\n* **MD5** \u2014 The base64-encoded MD5 hash of the object.\n\nFor an example of a valid TSV file, see\n[Transferring data from\nURLs](https://cloud.google.com/storage-transfer/docs/create-url-list).\n\nWhen transferring data based on a URL list, keep the following in mind:\n\n* When an object located at `http(s)://hostname:port/<URL-path>` is\ntransferred to a data sink, the name of the object at the data sink is\n`<hostname>/<URL-path>`.\n\n* If the specified size of an object does not match the actual size of the\nobject fetched, the object will not be transferred.\n\n* If the specified MD5 does not match the MD5 computed from the transferred\nbytes, the object transfer will fail. For more information, see\n[Generating MD5\nhashes](https://cloud.google.com/storage-transfer/docs/create-url-list#md5)\n\n* Ensure that each URL you specify is publicly accessible. For\nexample, in Cloud Storage you can\n[share an object publicly]\n(https://cloud.google.com/storage/docs/cloud-console#_sharingdata) and get\na link to it.\n\n* Storage Transfer Service obeys `robots.txt` rules and requires the source\nHTTP server to support `Range` requests and to return a `Content-Length`\nheader in each response.\n\n* ObjectConditions have no effect when filtering objects to transfer.",
"id": "HttpData",
"properties": {
"listUrl": {
@@ -629,31 +635,85 @@
},
"type": "object"
},
"NotificationConfig": {
"description": "Specification to configure notifications published to Cloud Pub/Sub.\nNotifications will be published to the customer-provided topic using the\nfollowing `PubsubMessage.attributes`:\n\n* `\"eventType\"`: one of the EventType values\n* `\"payloadFormat\"`: one of the PayloadFormat values\n* `\"projectId\"`: the project_id of the\n`TransferOperation`\n* `\"transferJobName\"`: the\ntransfer_job_name of the\n`TransferOperation`\n* `\"transferOperationName\"`: the name of the\n`TransferOperation`\n\nThe `PubsubMessage.data` will contain a TransferOperation resource\nformatted according to the specified `PayloadFormat`.",
"id": "NotificationConfig",
"properties": {
"eventTypes": {
"description": "Event types for which a notification is desired. If empty, send\nnotifications for all event types.",
"enumDescriptions": [
"Illegal value, to avoid allowing a default.",
"`TransferOperation` completed with status\nSUCCESS.",
"`TransferOperation` completed with status\nFAILED.",
"`TransferOperation` completed with status\nABORTED."
],
"items": {
"enum": [
"EVENT_TYPE_UNSPECIFIED",
"TRANSFER_OPERATION_SUCCESS",
"TRANSFER_OPERATION_FAILED",
"TRANSFER_OPERATION_ABORTED"
],
"type": "string"
},
"type": "array"
},
"payloadFormat": {
"description": "Required. The desired format of the notification message payloads.",
"enum": [
"PAYLOAD_FORMAT_UNSPECIFIED",
"NONE",
"JSON"
],
"enumDescriptions": [
"Illegal value, to avoid allowing a default.",
"No payload is included with the notification.",
"`TransferOperation` is [formatted as a JSON\nresponse](https://developers.google.com/protocol-buffers/docs/proto3#json),\nin application/json."
],
"type": "string"
},
"pubsubTopic": {
"description": "Required. The `Topic.name` of the Cloud Pub/Sub topic to which to publish\nnotifications. Must be of the format: `projects/{project}/topics/{topic}`.\nNot matching this format will result in an\nINVALID_ARGUMENT error.",
"type": "string"
}
},
"type": "object"
},
"ObjectConditions": {
"description": "Conditions that determine which objects will be transferred. Applies only\nto S3 and GCS objects.",
"description": "Conditions that determine which objects will be transferred. Applies only\nto S3 and Cloud Storage objects.\n\nThe \"last modification time\" refers to the time of the\nlast change to the object's content or metadata \u2014 specifically, this is\nthe `updated` property of Cloud Storage objects and the `LastModified`\nfield of S3 objects.",
"id": "ObjectConditions",
"properties": {
"excludePrefixes": {
"description": "`excludePrefixes` must follow the requirements described for\n`includePrefixes`.\n\nThe max size of `excludePrefixes` is 1000.",
"description": "`exclude_prefixes` must follow the requirements described for\ninclude_prefixes.\n\nThe max size of `exclude_prefixes` is 1000.",
"items": {
"type": "string"
},
"type": "array"
},
"includePrefixes": {
"description": "If `includePrefixes` is specified, objects that satisfy the object\nconditions must have names that start with one of the `includePrefixes`\nand that do not start with any of the `excludePrefixes`. If\n`includePrefixes` is not specified, all objects except those that have\nnames starting with one of the `excludePrefixes` must satisfy the object\nconditions.\n\nRequirements:\n\n * Each include-prefix and exclude-prefix can contain any sequence of\n Unicode characters, of max length 1024 bytes when UTF8-encoded, and\n must not contain Carriage Return or Line Feed characters. Wildcard\n matching and regular expression matching are not supported.\n\n * Each include-prefix and exclude-prefix must omit the leading slash.\n For example, to include the `requests.gz` object in a transfer from\n `s3://my-aws-bucket/logs/y=2015/requests.gz`, specify the include\n prefix as `logs/y=2015/requests.gz`.\n\n * None of the include-prefix or the exclude-prefix values can be empty,\n if specified.\n\n * Each include-prefix must include a distinct portion of the object\n namespace, i.e., no include-prefix may be a prefix of another\n include-prefix.\n\n * Each exclude-prefix must exclude a distinct portion of the object\n namespace, i.e., no exclude-prefix may be a prefix of another\n exclude-prefix.\n\n * If `includePrefixes` is specified, then each exclude-prefix must start\n with the value of a path explicitly included by `includePrefixes`.\n\nThe max size of `includePrefixes` is 1000.",
"description": "If `include_prefixes` is specified, objects that satisfy the object\nconditions must have names that start with one of the `include_prefixes`\nand that do not start with any of the exclude_prefixes. If\n`include_prefixes` is not specified, all objects except those that have\nnames starting with one of the `exclude_prefixes` must satisfy the object\nconditions.\n\nRequirements:\n\n * Each include-prefix and exclude-prefix can contain any sequence of\n Unicode characters, to a max length of 1024 bytes when UTF8-encoded,\n and must not contain Carriage Return or Line Feed characters. Wildcard\n matching and regular expression matching are not supported.\n\n * Each include-prefix and exclude-prefix must omit the leading slash.\n For example, to include the `requests.gz` object in a transfer from\n `s3://my-aws-bucket/logs/y=2015/requests.gz`, specify the include\n prefix as `logs/y=2015/requests.gz`.\n\n * None of the include-prefix or the exclude-prefix values can be empty,\n if specified.\n\n * Each include-prefix must include a distinct portion of the object\n namespace. No include-prefix may be a prefix of another\n include-prefix.\n\n * Each exclude-prefix must exclude a distinct portion of the object\n namespace. No exclude-prefix may be a prefix of another\n exclude-prefix.\n\n * If `include_prefixes` is specified, then each exclude-prefix must start\n with the value of a path explicitly included by `include_prefixes`.\n\nThe max size of `include_prefixes` is 1000.",
"items": {
"type": "string"
},
"type": "array"
},
"lastModifiedBefore": {
"description": "If specified, only objects with a \"last modification time\" before this\ntimestamp and objects that don't have a \"last modification time\" will be\ntransferred.",
"format": "google-datetime",
"type": "string"
},
"lastModifiedSince": {
"description": "If specified, only objects with a \"last modification time\" on or after\nthis timestamp and objects that don't have a \"last modification time\" are\ntransferred.\n\nThe `last_modified_since` and `last_modified_before` fields can be used\ntogether for chunked data processing. For example, consider a script that\nprocesses each day's worth of data at a time. For that you'd set each\nof the fields as follows:\n\n* `last_modified_since` to the start of the day\n\n* `last_modified_before` to the end of the day",
"format": "google-datetime",
"type": "string"
},
"maxTimeElapsedSinceLastModification": {
"description": "If specified, only objects with a `lastModificationTime` on or after\n`NOW` - `maxTimeElapsedSinceLastModification` and objects that don't have\na `lastModificationTime` are transferred.\n\nNote that, for each `TransferOperation` started by this `TransferJob`,\n`NOW` refers to the `start_time` of the 'TransferOperation`. Also,\n`lastModificationTime` refers to the time of the last change to the\nobject's content or metadata - specifically, this would be the `updated`\nproperty of GCS objects and the `LastModified` field of S3 objects.",
"description": "If specified, only objects with a \"last modification time\" on or after\n`NOW` - `max_time_elapsed_since_last_modification` and objects that don't\nhave a \"last modification time\" are transferred.\n\nFor each TransferOperation started by this TransferJob,\n`NOW` refers to the start_time of the\n`TransferOperation`.",
"format": "google-duration",
"type": "string"
},
"minTimeElapsedSinceLastModification": {
"description": "If specified, only objects with a `lastModificationTime` before\n`NOW` - `minTimeElapsedSinceLastModification` and objects that don't have a\n`lastModificationTime` are transferred.\n\nNote that, for each `TransferOperation` started by this `TransferJob`,\n`NOW` refers to the `start_time` of the 'TransferOperation`. Also,\n`lastModificationTime` refers to the time of the last change to the\nobject's content or metadata - specifically, this would be the `updated`\nproperty of GCS objects and the `LastModified` field of S3 objects.",
"description": "If specified, only objects with a \"last modification time\" before\n`NOW` - `min_time_elapsed_since_last_modification` and objects that don't\n have a \"last modification time\" are transferred.\n\nFor each TransferOperation started by this TransferJob, `NOW`\nrefers to the start_time of the\n`TransferOperation`.",
"format": "google-duration",
"type": "string"
}
@@ -713,15 +773,15 @@
"properties": {
"scheduleEndDate": {
"$ref": "Date",
"description": "The last day the recurring transfer will be run. If `scheduleEndDate`\nis the same as `scheduleStartDate`, the transfer will be executed only\nonce."
"description": "The last day a transfer runs. Date boundaries are determined relative to\nUTC time. A job will run once per 24 hours within the following guidelines:\n\n* If `schedule_end_date` and schedule_start_date are the same and in\n the future relative to UTC, the transfer is executed only one time.\n* If `schedule_end_date` is later than `schedule_start_date` and\n `schedule_end_date` is in the future relative to UTC, the job will\n run each day at start_time_of_day through `schedule_end_date`."
},
"scheduleStartDate": {
"$ref": "Date",
"description": "Required. The first day the recurring transfer is scheduled to run. If\n`scheduleStartDate` is in the past, the transfer will run for the first\ntime on the following day."
"description": "Required. The start date of a transfer. Date boundaries are determined\nrelative to UTC time. If `schedule_start_date` and start_time_of_day\nare in the past relative to the job's creation time, the transfer starts\nthe day after you schedule the transfer request.\n\n**Note:** When starting jobs at or near midnight UTC it is possible that\na job will start later than expected. For example, if you send an outbound\nrequest on June 1 one millisecond prior to midnight UTC and the Storage\nTransfer Service server receives the request on June 2, then it will create\na TransferJob with `schedule_start_date` set to June 2 and a\n`start_time_of_day` set to midnight UTC. The first scheduled\nTransferOperation will take place on June 3 at midnight UTC."
},
"startTimeOfDay": {
"$ref": "TimeOfDay",
"description": "The time in UTC at which the transfer will be scheduled to start in a day.\nTransfers may start later than this time. If not specified, recurring and\none-time transfers that are scheduled to run today will run immediately;\nrecurring transfers that are scheduled to run on a future date will start\nat approximately midnight UTC on that date. Note that when configuring a\ntransfer with the Cloud Platform Console, the transfer's start time in a\nday is specified in your local timezone."
"description": "The time in UTC that a transfer job is scheduled to run. Transfers may\nstart later than this time.\n\nIf `start_time_of_day` is not specified:\n\n* One-time transfers run immediately.\n* Recurring transfers run immediately, and each day at midnight UTC,\n through schedule_end_date.\n\nIf `start_time_of_day` is specified:\n\n* One-time transfers run at the specified time.\n* Recurring transfers run at the specified time each day, through\n `schedule_end_date`."
}
},
"type": "object"
@@ -891,9 +951,13 @@
"type": "string"
},
"name": {
"description": "A globally unique name assigned by Storage Transfer Service when the\njob is created. This field should be left empty in requests to create a new\ntransfer job; otherwise, the requests result in an `INVALID_ARGUMENT`\nerror.",
"description": "A unique name (within the transfer project) assigned when the job is\ncreated. If this field is empty in a CreateTransferJobRequest, Storage\nTransfer Service will assign a unique name. Otherwise, the specified name\nis used as the unique name for this job.\n\nIf the specified name is in use by a job, the creation request fails with\nan ALREADY_EXISTS error.\n\nThis name must start with `\"transferJobs/\"` prefix and end with a letter or\na number, and should be no more than 128 characters.\nExample: `\"transferJobs/[A-Za-z0-9-._~]*[A-Za-z0-9]$\"`\n\nInvalid job names will fail with an\nINVALID_ARGUMENT error.",
"type": "string"
},
"notificationConfig": {
"$ref": "NotificationConfig",
"description": "Notification configuration."
},
"projectId": {
"description": "The ID of the Google Cloud Platform Project that owns the job.",
"type": "string"
@@ -903,7 +967,7 @@
"description": "Schedule specification."
},
"status": {
"description": "Status of the job. This value MUST be specified for\n`CreateTransferJobRequests`.\n\nNOTE: The effect of the new job status takes place during a subsequent job\nrun. For example, if you change the job status from `ENABLED` to\n`DISABLED`, and an operation spawned by the transfer is running, the status\nchange would not affect the current operation.",
"description": "Status of the job. This value MUST be specified for\n`CreateTransferJobRequests`.\n\n**Note:** The effect of the new job status takes place during a subsequent\njob run. For example, if you change the job status from\nENABLED to DISABLED, and an operation\nspawned by the transfer is running, the status change would not affect the\ncurrent operation.",
"enum": [
"STATUS_UNSPECIFIED",
"ENABLED",
@@ -949,6 +1013,10 @@
"description": "A globally unique ID assigned by the system.",
"type": "string"
},
"notificationConfig": {
"$ref": "NotificationConfig",
"description": "Notification configuration."
},
"projectId": {
"description": "The ID of the Google Cloud Platform Project that owns the operation.",
"type": "string"
@@ -966,7 +1034,8 @@
"PAUSED",
"SUCCESS",
"FAILED",
"ABORTED"
"ABORTED",
"QUEUED"
],
"enumDescriptions": [
"Zero is an illegal value.",
@@ -974,7 +1043,8 @@
"Paused.",
"Completed successfully.",
"Terminated due to an unrecoverable failure.",
"Aborted by the user."
"Aborted by the user.",
"Temporarily delayed by the system. No user action is required."
],
"type": "string"
},
@@ -994,11 +1064,11 @@
"id": "TransferOptions",
"properties": {
"deleteObjectsFromSourceAfterTransfer": {
"description": "Whether objects should be deleted from the source after they are\ntransferred to the sink. Note that this option and\n`deleteObjectsUniqueInSink` are mutually exclusive.",
"description": "Whether objects should be deleted from the source after they are\ntransferred to the sink.\n\n**Note:** This option and delete_objects_unique_in_sink are mutually\nexclusive.",
"type": "boolean"
},
"deleteObjectsUniqueInSink": {
"description": "Whether objects that exist only in the sink should be deleted. Note that\nthis option and `deleteObjectsFromSourceAfterTransfer` are mutually\nexclusive.",
"description": "Whether objects that exist only in the sink should be deleted.\n\n**Note:** This option and delete_objects_from_source_after_transfer are\nmutually exclusive.",
"type": "boolean"
},
"overwriteObjectsAlreadyExistingInSink": {
@@ -1016,13 +1086,17 @@
"$ref": "AwsS3Data",
"description": "An AWS S3 data source."
},
"azureBlobStorageDataSource": {
"$ref": "AzureBlobStorageData",
"description": "An Azure Blob Storage data source."
},
"gcsDataSink": {
"$ref": "GcsData",
"description": "A Google Cloud Storage data sink."
"description": "A Cloud Storage data sink."
},
"gcsDataSource": {
"$ref": "GcsData",
"description": "A Google Cloud Storage data source."
"description": "A Cloud Storage data source."
},
"httpDataSource": {
"$ref": "HttpData",
@@ -1030,11 +1104,11 @@
},
"objectConditions": {
"$ref": "ObjectConditions",
"description": "Only objects that satisfy these object conditions are included in the set\nof data source and data sink objects. Object conditions based on\nobjects' `lastModificationTime` do not exclude objects in a data sink."
"description": "Only objects that satisfy these object conditions are included in the set\nof data source and data sink objects. Object conditions based on\nobjects' \"last modification time\" do not exclude objects in a data sink."
},
"transferOptions": {
"$ref": "TransferOptions",
"description": "If the option `deleteObjectsUniqueInSink` is `true`, object conditions\nbased on objects' `lastModificationTime` are ignored and do not exclude\nobjects in a data source or a data sink."
"description": "If the option\ndelete_objects_unique_in_sink\nis `true`, object conditions based on objects' \"last modification time\" are\nignored and do not exclude objects in a data source or a data sink."
}
},
"type": "object"
@@ -1049,10 +1123,10 @@
},
"transferJob": {
"$ref": "TransferJob",
"description": "Required. The job to update. `transferJob` is expected to specify only\nthree fields: `description`, `transferSpec`, and `status`. An\nUpdateTransferJobRequest that specifies other fields will be rejected with\nan error `INVALID_ARGUMENT`."
"description": "Required. The job to update. `transferJob` is expected to specify only\nfour fields: description,\ntransfer_spec,\nnotification_config, and\nstatus. An `UpdateTransferJobRequest` that specifies\nother fields will be rejected with the error\nINVALID_ARGUMENT."
},
"updateTransferJobFieldMask": {
"description": "The field mask of the fields in `transferJob` that are to be updated in\nthis request. Fields in `transferJob` that can be updated are:\n`description`, `transferSpec`, and `status`. To update the `transferSpec`\nof the job, a complete transfer specification has to be provided. An\nincomplete specification which misses any required fields will be rejected\nwith the error `INVALID_ARGUMENT`.",
"description": "The field mask of the fields in `transferJob` that are to be updated in\nthis request. Fields in `transferJob` that can be updated are:\ndescription,\ntransfer_spec,\nnotification_config, and\nstatus. To update the `transfer_spec` of the job, a\ncomplete transfer specification must be provided. An incomplete\nspecification missing any required fields will be rejected with the error\nINVALID_ARGUMENT.",
"format": "google-fieldmask",
"type": "string"
}