Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support S3 for cache files #849

Merged

Conversation

Mr0grog
Copy link
Member

@Mr0grog Mr0grog commented Feb 16, 2023

To support running this job on an actual scheduled job runner that doesn't have persistent storage (see #757), we need to be able to store the unplaybackable cache in S3. You can now use 's3://' paths in the --unplaybackable option:

wm import ia 'https://somewhere.com/' --unplaybackable 's3://bucket/unplaybackable.json'

To support running this job on an actual scheduled job runner that doesn't have persistent storage, we need to be able to store the unplaybackable cache in S3. You can now use 's3://' paths in the `--unplaybackable` option:

    wm import ia 'https://somewhere.com/' --unplaybackable 's3://bucket/unplaybackable.json'
@Mr0grog Mr0grog merged commit 250c554 into main Feb 16, 2023
@Mr0grog Mr0grog deleted the sometimes-the-local-filesystem-is-not-a-very-good-cache branch February 16, 2023 17:54
Mr0grog added a commit that referenced this pull request Feb 16, 2023
Mr0grog added a commit to edgi-govdata-archiving/web-monitoring-ops that referenced this pull request Feb 16, 2023
Instead of running the import job as a cron script on a random EC2 VM, run it as an actual CronJob in Kubernetes with everything else. This also cleans up the docs around jobs.

Work not visible here: created a new IAM account for jobs that can write to relevant S3 buckets, added ability to store cache files in S3 (edgi-govdata-archiving/web-monitoring-processing#849) since we have no persistent storage in Kubernetes.

Why do this now? See:
- edgi-govdata-archiving/web-monitoring#168
- edgi-govdata-archiving/web-monitoring-processing#757
Mr0grog added a commit to edgi-govdata-archiving/web-monitoring-ops that referenced this pull request Feb 16, 2023
Instead of running the import job as a cron script on a random EC2 VM, run it as an actual CronJob in Kubernetes with everything else. This also cleans up the docs around jobs.

Work not visible here: created a new IAM account for jobs that can write to relevant S3 buckets, added ability to store cache files in S3 (edgi-govdata-archiving/web-monitoring-processing#849) since we have no persistent storage in Kubernetes.

Why do this now? See:
- edgi-govdata-archiving/web-monitoring#168
- edgi-govdata-archiving/web-monitoring-processing#757
Mr0grog added a commit to edgi-govdata-archiving/web-monitoring-ops that referenced this pull request Feb 17, 2023
Instead of running the import job as a cron script on a random EC2 VM, run it as an actual CronJob in Kubernetes with everything else. This also cleans up the docs around jobs.

Why do this now? See:
- edgi-govdata-archiving/web-monitoring#168
- edgi-govdata-archiving/web-monitoring-processing#757

Work not visible here:
- Created a new IAM account for jobs that can write to relevant S3 buckets.
- Added ability to store cache files in S3 (edgi-govdata-archiving/web-monitoring-processing#849) since we have no persistent storage in Kubernetes.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant