File: //snap/google-cloud-cli/current/lib/surface/transfer/jobs/create.py
# -*- coding: utf-8 -*- #
# Copyright 2021 Google LLC. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Command to create transfer jobs."""
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
from googlecloudsdk.api_lib.transfer import operations_util
from googlecloudsdk.api_lib.util import apis
from googlecloudsdk.calliope import base
from googlecloudsdk.command_lib.storage import storage_url
from googlecloudsdk.command_lib.transfer import jobs_apitools_util
from googlecloudsdk.command_lib.transfer import jobs_flag_util
from googlecloudsdk.core import log
@base.UniverseCompatible
class Create(base.Command):
"""Create a Transfer Service transfer job."""
# pylint:disable=line-too-long
detailed_help = {
'DESCRIPTION':
"""\
Create a Transfer Service transfer job, allowing you to transfer data to
Google Cloud Storage on a one-time or recurring basis.
""",
'EXAMPLES':
"""\
To create a one-time, immediate transfer job to move data from Google
Cloud Storage bucket "foo" into the "baz" folder in Cloud Storage bucket
"bar", run:
$ {command} gs://foo gs://bar/baz/
To create a transfer job to move data from an Amazon S3 bucket called
"foo" to a Google Cloud Storage bucket named "bar" that runs every day
with custom name "my-test-job", run:
$ {command} s3://foo gs://bar --name=my-test-job --source-creds-file=/examplefolder/creds.txt --schedule-repeats-every=1d
To create a one-time, immediate transfer job to move data between Google
Cloud Storage buckets "foo" and "bar" with filters to include objects that
start with prefixes "baz" and "qux"; and objects modified in the 24 hours
before the transfer started, run:
$ {command} gs://foo gs://bar/ --include-prefixes=baz,qux --include-modified-after-relative=1d
To create a one-time, immediate transfer job to move data from a directory
with absolute path "/foo/bar/" in the filesystem associated with
agent pool "my-pool" into Google Cloud Storage bucket "example-bucket",
run:
$ {command} posix:///foo/bar/ gs://example-bucket --source-agent-pool=my-pool
"""
}
# pylint:enable=line-too-long
@classmethod
def Args(cls, parser):
jobs_flag_util.setup_parser(parser, release_track=cls.ReleaseTrack())
def Run(self, args):
is_hdfs_source = args.source.startswith(
storage_url.ProviderPrefix.HDFS.value
)
is_posix_source = args.source.startswith(
storage_url.ProviderPrefix.POSIX.value
)
is_posix_destination = args.destination.startswith(
storage_url.ProviderPrefix.POSIX.value
)
if (is_hdfs_source or is_posix_source) and not args.source_agent_pool:
raise ValueError(
'Missing agent pool. Please add --source-agent-pool flag.')
if is_posix_destination and not args.destination_agent_pool:
raise ValueError(
'Missing agent pool. Please add --destination-agent-pool flag.')
if (is_posix_source and is_posix_destination and
not args.intermediate_storage_path):
raise ValueError('Missing intermediate storage path.'
' Please add --intermediate-storage-path flag.')
client = apis.GetClientInstance('transfer', 'v1')
messages = apis.GetMessagesModule('transfer', 'v1')
result = client.transferJobs.Create(
jobs_apitools_util.generate_transfer_job_message(args, messages))
if args.no_async:
log.status.Print('Created job: {}'.format(result.name))
operations_util.block_until_done(job_name=result.name)
return result