Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix minor typo in docs/cloud-integration.md #21629

Closed
wants to merge 1 commit into from
Closed
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 3 additions & 2 deletions docs/cloud-integration.md
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ be safely used as the direct destination of work with the normal rename-based co
### Installation

With the relevant libraries on the classpath and Spark configured with valid credentials,
objects can be can be read or written by using their URLs as the path to data.
objects can be read or written by using their URLs as the path to data.
For example `sparkContext.textFile("s3a://landsat-pds/scene_list.gz")` will create
an RDD of the file `scene_list.gz` stored in S3, using the s3a connector.

Expand Down Expand Up @@ -184,7 +184,8 @@ is no need for a workflow of write-then-rename to ensure that files aren't picke
while they are still being written. Applications can write straight to the monitored directory.

1. Streams should only be checkpointed to a store implementing a fast and
atomic `rename()` operation Otherwise the checkpointing may be slow and potentially unreliable.
atomic `rename()` operation.
Otherwise the checkpointing may be slow and potentially unreliable.

## Further Reading

Expand Down