From cdbc1d172c00c30a4579c98bd38d51e0690330c9 Mon Sep 17 00:00:00 2001 From: Bart Schuurmans Date: Wed, 27 Mar 2024 23:27:19 +0100 Subject: [PATCH] Fix double exports subdir in S3 user export --- bookwyrm/models/bookwyrm_export_job.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/bookwyrm/models/bookwyrm_export_job.py b/bookwyrm/models/bookwyrm_export_job.py index 7a0c1100c..35226c6a4 100644 --- a/bookwyrm/models/bookwyrm_export_job.py +++ b/bookwyrm/models/bookwyrm_export_job.py @@ -99,6 +99,7 @@ def create_archive_task(job_id): try: export_task_id = job.task_id + archive_filename = f"{export_task_id}.tar.gz" export_json_bytes = DjangoJSONEncoder().encode(job.export_json).encode("utf-8") user = job.user @@ -109,10 +110,9 @@ def create_archive_task(job_id): storage = S3Boto3Storage() # Handle for creating the final archive - s3_archive_path = f"exports/{export_task_id}.tar.gz" s3_tar = S3Tar( settings.AWS_STORAGE_BUCKET_NAME, - s3_archive_path, + f"exports/{archive_filename}", session=BookwyrmAwsSession(), ) @@ -135,14 +135,14 @@ def create_archive_task(job_id): # Create archive and store file name s3_tar.tar() - job.export_data = s3_archive_path + job.export_data = archive_filename job.save(update_fields=["export_data"]) # Delete temporary files S3Boto3Storage.delete(storage, export_json_tmp_file) else: - job.export_data = f"{export_task_id}.tar.gz" + job.export_data = archive_filename with job.export_data.open("wb") as tar_file: with BookwyrmTarFile.open(mode="w:gz", fileobj=tar_file) as tar: # save json file