-
Notifications
You must be signed in to change notification settings - Fork 122
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Write Items or collections in parallel #690
Comments
@ljstrnadiii Thanks for raising this. I opened #749 to add async I/O operations to the library, which should help with this.
The answer to this really depends on the content and structure of your GeoTIFFs. STAC Items are meant to represent distinct spatiotemporal resources (think satellite imagery scenes), so and files that are part of the same scene should be represented as assets in a single Item. However, if the GeoTIFFs are not part of a single scene then you are probably doing the right thing by creating separate Items. @m-mohr @cholmes May be able to offer additional guidance on this as well.
I've opened #749 to introduce async I/O operations into the library, and any feedback on that would be much appreciated. Currently, the recursive saving of Items happens here. In the PR, those save requests are made in batches of asynchronous requests here.
Calling
...and PySTAC will only save the root Catalog and the modified/new Collection without walking the rest of the Catalog. |
We discussed |
I am pretty new to stac and pystac, but am currently giving it a try.
I have 1tb of geotiffs and I have build a function to build items by scanning over all elements, categorizing into collections, adding collections and writing out to a catalog. I wrote a custom stackio object for a first pass, simple way to persist to gcs:
as described here.
When I save out, it takes some time since we seem to be saving out many very small jsons (one per item). Is there any way to write these out in parallel? (assuming each item should in fact be saved out as a small json element)
So, two questions:
Thanks a ton! Cool stuff!
The text was updated successfully, but these errors were encountered: