+1 to all of those conventions. And although that is a lot of classes in airflow/providers/aws/operators/s3.py I like it that way rather than split out.
The only odd on is obviously `lambda_` but we have to do something to that. -ash > On 30 Oct 2019, at 08:13, Bas Harenslak <basharens...@godatadriven.com> wrote: > > Hi, > > How about the following conventions? > > * No AWS in whatever notation in the class name, since this is already in > the module (airflow.providers.aws….) > * Only exception to above is the AwsHook, which is a “base” hook which > most other hooks inherit from > * No “Transfer” in class names, since all transfer operators already have > “to” in the name, e.g. RedshiftToS3Transfer becomes RedshiftToS3Operator > * For all class names, the naming convention is {service}{component}, e.g. > “AthenaHook" > * For all “transfer” names, the naming convention is > {service}To{service}{component}, e.g. “RedshiftToS3Operator" > * Migrate all AWS related components to /airflow/providers/aws/… > * Sidenote: lambda is a reserved Python keyword, so name the Lambda > service files "lambda_.py", according to Python conventions to avoid name > clashes > > Applying these rules, I get the following changes (done some regex, might > have missed a few): > > ./contrib/hooks/aws_dynamodb_hook.py:AwsDynamoDBHook -> > ./providers/aws/hooks/dynamodb.py:DynamoDBHook > ./contrib/hooks/aws_firehose_hook.py:AwsFirehoseHook -> > ./providers/aws/hooks/firehose.py:FirehoseHook > ./contrib/hooks/aws_glue_catalog_hook.py:AwsGlueCatalogHook -> > ./providers/aws/hooks/glue.py:GlueHook > ./contrib/hooks/aws_hook.py:AwsHook -> > ./providers/aws/hooks/base/base.py:AwsHook > ./contrib/hooks/aws_lambda_hook.py:AwsLambdaHook -> > ./providers/aws/hooks/lambda_.py:LambdaHook > ./contrib/hooks/aws_logs_hook.py:AwsLogsHook -> > ./providers/aws/hooks/cloudwatch.py:CloudwatchLogsHook > ./contrib/hooks/aws_sns_hook.py:AwsSnsHook -> > ./providers/aws/hooks/sns.py:SnsHook > ./contrib/hooks/aws_sqs_hook.py:SQSHook -> > ./providers/aws/hooks/sqs.py:SqsHook > ./contrib/hooks/emr_hook.py:EmrHook -> ./providers/aws/hooks/emr.py:EmrHook > ./contrib/hooks/redshift_hook.py:RedshiftHook -> > ./providers/aws/hooks/redshift.py:RedshiftHook > ./contrib/hooks/sagemaker_hook.py:SageMakerHook -> > ./providers/aws/hooks/sagemaker.py:SagemakerHook > ./hooks/S3_hook.py:S3Hook -> ./providers/aws/hooks/s3.py:S3Hook > ./providers/aws/hooks/athena.py:AWSAthenaHook -> > ./providers/aws/hooks/athena.py:AthenaHook > > ./contrib/operators/aws_sqs_publish_operator.py:SQSPublishOperator -> > ./providers/aws/operators/sqs.py:SQSPublishOperator > ./contrib/operators/dynamodb_to_s3.py:DynamoDBToS3Operator -> > ./providers/aws/operators/s3.py:DynamoDBToS3Operator > ./contrib/operators/hive_to_dynamodb.py:HiveToDynamoDBTransferOperator -> > ./providers/aws/operators/dynamodb.py:HiveToDynamoDBOperator > ./contrib/operators/imap_attachment_to_s3_operator.py:ImapAttachmentToS3Operator > -> ./providers/aws/operators/s3.py:ImapAttachmentToS3Operator > ./contrib/operators/mongo_to_s3.py:MongoToS3Operator -> > ./providers/aws/operators/s3.py:MongoToS3Operator > ./contrib/operators/s3_copy_object_operator.py:S3CopyObjectOperator -> > ./providers/aws/operators/s3.py:S3CopyObjectOperator > ./contrib/operators/s3_delete_objects_operator.py:S3DeleteObjectsOperator -> > ./providers/aws/operators/s3.py:S3DeleteObjectsOperator > ./contrib/operators/s3_list_operator.py:S3ListOperator -> > ./providers/aws/operators/s3.py:S3ListOperator > ./contrib/operators/s3_to_gcs_operator.py:S3ToGoogleCloudStorageOperator -> > ./providers/aws/operators/s3.py:S3ToGoogleCloudStorageOperator > ./contrib/operators/s3_to_sftp_operator.py:S3ToSFTPOperator -> > ./providers/aws/operators/s3.py:S3ToSFTPOperator > ./contrib/operators/sftp_to_s3_operator.py:SFTPToS3Operator -> > ./providers/aws/operators/s3.py:SFTPToS3Operator > ./operators/gcs_to_s3.py:GoogleCloudStorageToS3Operator -> > ./providers/aws/operators/s3.py:GoogleCloudStorageToS3Operator > ./operators/google_api_to_s3_transfer.py:GoogleApiToS3Transfer -> > ./providers/aws/operators/s3.py:GoogleApiToS3Operator > ./operators/redshift_to_s3_operator.py:RedshiftToS3Transfer -> > ./providers/aws/operators/s3.py:RedshiftToS3Operator > ./operators/s3_file_transform_operator.py:S3FileTransformOperator -> > ./providers/aws/operators/s3.py:S3FileTransformOperator > ./operators/s3_to_hive_operator.py:S3ToHiveTransfer -> > ./providers/aws/operators/s3.py:S3ToHiveOperator > ./operators/s3_to_redshift_operator.py:S3ToRedshiftTransfer -> > ./providers/aws/operators/s3.py:S3ToRedshiftOperator > ./providers/aws/operators/athena.py:AWSAthenaOperator -> > ./providers/aws/operators/athena.py:AthenaOperator > ./providers/aws/operators/batch.py:AWSBatchOperator -> > ./providers/aws/operators/batch.py:BatchOperator > > ./contrib/sensors/aws_glue_catalog_partition_sensor.py:AwsGlueCatalogPartitionSensor > -> ./providers/aws/sensors/glue.py:GlueCatalogPartitionSensor > ./contrib/sensors/aws_redshift_cluster_sensor.py:AwsRedshiftClusterSensor -> > ./providers/aws/sensors/redshift.py:RedshiftClusterSensor > ./contrib/sensors/aws_sqs_sensor.py:SQSSensor -> > ./providers/aws/sensors/sqs.py:SQSSensor > ./providers/aws/sensors/athena.py:AthenaSensor -> > ./providers/aws/sensors/athena.py:AthenaSensor > ./sensors/s3_key_sensor.py:S3KeySensor -> > ./providers/aws/sensors/s3.py:S3KeySensor > ./sensors/s3_prefix_sensor.py:S3PrefixSensor -> > ./providers/aws/sensors/s3.py:S3PrefixSensor > > WDYT? > > Bas > > On 30 Oct 2019, at 08:52, Tomasz Urbaszek > <tomasz.urbas...@polidea.com<mailto:tomasz.urbas...@polidea.com>> wrote: > > I think it's a great idea and I am willing to help. > > Bests, > Tomek > > On Wed, Oct 30, 2019 at 3:17 AM MinJae Kwon > <mingram...@gmail.com<mailto:mingram...@gmail.com>> wrote: > > Hi all. > > While I was working on AIRFLOW-5803, I found there are inconsistencies > between class names for each AWS operators/hooks. > > For example, Athena hook is named as 'AWSAthenaHook', but Dynamodb hook is > named as 'AwsDynamoDBHook', and S3 hook is just 'S3Hook' (So, AIRFLOW-5803 > is trying to rename S3Hook to AWSS3Hook) > > So we might need to have some conventions or guidelines for AWS > integrations like GCP ( > https://docs.google.com/document/d/1_rTdJSLCt0eyrAylmmgYc3yZr-_h51fVlnvMmWqhCkY/edit?ts=5bb72dfd# > ) > > What do you think about it? > > Thank you. > > > > -- > > Tomasz Urbaszek > Polidea <https://www.polidea.com/> | Junior Software Engineer > > M: +48 505 628 493 <+48505628493> > E: tomasz.urbas...@polidea.com<mailto:tomasz.urbas...@polidea.com> > <tomasz.urbasz...@polidea.com<mailto:tomasz.urbasz...@polidea.com>> > > Unique Tech > Check out our projects! <https://www.polidea.com/our-work> >