See the documentation
for setting up your S3 API keys, especially the
Configuration Settings and Precedence section.
Typically use environment variables or the AWS credentials file.
Neither requires any changes in the code.
Simply set your API keys with your method of choice
and start using the program.
This application assumes that at least one bucket has been
created in the AWS Console; it will list all available buckets
that are accessible with the S3 Access/Secret keys being used.
A trailing /
after a name in the file explorer denotes a directory.
A local directory may only be uploaded to S3 using s3zilla if it
contains at least one file. S3 itself doesn't really have a concept
of folders, but s3zilla directory uploads will follow the same
structure as the given folder being uploaded. Therefor directories
found in an S3 bucket that are visible within s3zilla are not actually
directories; they are a representation of directories that are
actually S3 file objects.
For example, if a directory is uploaded to
S3 from your local machine with the name test/
and contains two files
named test1.pdf
and test2.docx
, they will appear as test/test1.pdf
and test/test2.docx
in S3.
Likewise if an S3 object named test/some_folder/file.txt
is downloaded
from S3 to the local machine, a folder named test/
will be placed
in the chosen local directory. It will contain another folder named
some_folder
that contains file.txt
. These directories will be created
if they do not already exist.
s3zilla does not currently have the capability to set
private/public/readOnly ACLs on uploaded file-objects; AWS now disables
this feature by default when creating a new bucket.
Multiple files can be selected at once when uploading, downloading, and/or deleting
local files and remote S3 objects.
Requirements:
pip install boto3
Tested on Windows 10/11 Author: rootVIII 2019-2023