diff --git a/src/scripts/backfill_jobs/backfill_verifications_athena.sql b/src/scripts/backfill_jobs/backfill_verifications_athena.sql new file mode 100644 index 0000000..c5a743c --- /dev/null +++ b/src/scripts/backfill_jobs/backfill_verifications_athena.sql @@ -0,0 +1,152 @@ +CREATE EXTERNAL TABLE backfill.old_verifications ( + change_timestamp bigint, + record_type string, + json_record string +) +PARTITIONED BY (instance string, snapshot_date string) +ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.OpenCSVSerde' +WITH SERDEPROPERTIES ( "separatorChar"=",") +LOCATION 's3://prod.snapshot.record.sagebase.org/' +TBLPROPERTIES ( + "storage.location.template"="s3://prod.snapshot.record.sagebase.org/${instance}/verificationsubmission/${snapshot_date}/", + "projection.enabled"="true", + "projection.instance.type"="integer", + "projection.instance.range"="000000386,000000458", + "projection.instance.digits"="9", + "projection.snapshot_date.type"="date", + "projection.snapshot_date.format"="yyyy-MM-dd", + "projection.snapshot_date.range"="2022-01-01,2023-06-28" +); + +CREATE EXTERNAL TABLE backfill.transformed_verifications( + `stack` string, + `instance` string, + `objecttype` string, + `changetype` string, + `changetimestamp` bigint, + `snapshottimestamp` bigint, + `userid` bigint, + `snapshot` struct< + id: string, + createdOn: bigint, + createdBy: string, + firstName: string, + lastName: string, + emails: array, + notificationEmail: string, + location: string, + company: string, + orcid: string, + stateHistory: array< + struct< + createdOn: bigint, + createdBy: string, + state: string, + reason: string, + notes: string + > + >, + attachments: array< + struct< + fileName: string, + id: string + > + > + > +) +PARTITIONED BY ( + `year` string, + `month` string, + `day` string) +ROW FORMAT SERDE + 'org.apache.hive.hcatalog.data.JsonSerDe' +STORED AS INPUTFORMAT + 'org.apache.hadoop.mapred.TextInputFormat' +OUTPUTFORMAT + 'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +LOCATION + 's3://prod.log.sagebase.org/verificationSubmissionSnapshots/records_backfill/' +TBLPROPERTIES ( + 'write.compression'='GZIP' +); + +INSERT INTO backfill.transformed_userprofile +SELECT 'prod' as stack, + -- Gets rid of the zero padding + cast(cast(instance as integer) as varchar) as instance, + 'VERIFICATION_SUBMISSION' as objectType, + -- We do not have the changeType in the old data, if the timstamp of the change is more than a second after the creation date treat it as an UPDATE + IF (date_diff('millisecond', from_unixtime(change_timestamp / 1000), from_iso8601_timestamp(json_extract_scalar(json_record, '$.createdOn'))) > 1000, 'UPDATE', 'CREATE') as changeType, + change_timestamp as changeTimestamp, + cast(to_unixtime(CAST(snapshot_date AS TIMESTAMP)) * 1000 as bigint) as snapshotTimestamp, + -- We do not have the userId of the change in the old data warehouse, uses the ownerId from the snapshot + cast(json_extract_scalar(json_record, '$.createdBy') as bigint) as userId, + CAST( + ROW( + json_extract_scalar(json_record, '$.id'), + cast(to_unixtime(from_iso8601_timestamp(json_extract_scalar(json_record, '$.createdOn'))) * 1000 as bigint), + json_extract_scalar(json_record, '$.createdBy'), + json_extract_scalar(json_record, '$.firstName'), + json_extract_scalar(json_record, '$.lastName'), + json_parse(json_query(json_record, 'lax $.emails')), + json_extract_scalar(json_record, '$.notificationEmail'), + json_extract_scalar(json_record, '$.location'), + json_extract_scalar(json_record, '$.company'), + json_extract_scalar(json_record, '$.orcid'), + -- The state history array contains the createdOn as a ISO timestamp string, since we need to convert it to a bigint we have to extract the structure and apply a transformation to each element + transform( + -- We first extract the state history as an array of json + cast(json_parse(json_query(json_record, 'lax $.stateHistory')) as array), + -- We now apply a transformation to each element, defining the structure of the json element + e -> CAST( + ROW( + cast(to_unixtime(from_iso8601_timestamp(json_extract_scalar(e, '$.createdOn'))) * 1000 as bigint), + json_extract_scalar(e, '$.createdBy'), + json_extract_scalar(e, '$.state'), + json_extract_scalar(e, '$.reason'), + json_extract_scalar(e, '$.notes') + ) AS ROW (createdOn bigint, createdBy varchar, state varchar, reason varchar, notes varchar) + ) + ), + json_parse(json_query(json_record, 'lax $.attachments')) + ) AS ROW( + id varchar, + createdOn bigint, + createdBy varchar, + firstName varchar, + lastName varchar, + emails ARRAY(varchar), + notificationEmail varchar, + location varchar, + company varchar, + orcid varchar, + stateHistory ARRAY( + ROW( + createdOn bigint, + createdBy varchar, + state varchar, + reason varchar, + notes varchar + ) + ), + attachments ARRAY( + ROW( + fileName varchar, + id varchar + ) + ) + )) AS snapshot, + cast(year(date(snapshot_date)) as varchar) as year, + -- Zero pads month and day + lpad(cast(month(date(snapshot_date)) as varchar),2,'0') as month, + lpad(cast(day(date(snapshot_date)) as varchar),2,'0') as day +FROM old_verifications +WHERE instance IN +('000000386', '000000387', '000000388', '000000389', '000000390', '000000391', '000000392', '000000393', '000000394', '000000395'); +-- ('000000396', '000000397', '000000398', '000000399', '000000400', '000000401', '000000402', '000000403', '000000404', '000000405'); +-- ('000000406', '000000407', '000000408', '000000409', '000000410', '000000411', '000000412', '000000413', '000000414', '000000415'); +-- ('000000416', '000000417', '000000418', '000000419', '000000420', '000000421', '000000422', '000000423', '000000424', '000000425'); +-- ('000000426', '000000427', '000000428', '000000429', '000000430', '000000431', '000000432', '000000433', '000000434', '000000435'); +-- ('000000436', '000000437', '000000438', '000000439', '000000440', '000000441', '000000442', '000000443', '000000444', '000000445'); +-- ('000000446', '000000447', '000000448', '000000449', '000000450', '000000451', '000000452', '000000453', '000000454', '000000455'); +-- ('000000456', '000000457', '000000458'); \ No newline at end of file