Spark-Hadoop Unix Installer
This package uses Python to download and unpack a pre-built version of Spark/Hadoop from Apache. Its primary use-case is simplifying unattended installs where the user wants "the latest available version" of these tools.
- download Spark/Hadoop release tarball from Apache.
- verify the tarball using the SHA512 sum provided by Apache.
- unpack the tarball to a target directory on your local system.
First you'll need to install shui
using pip: pip install shui
.
The versions
command shows you all available Spark/Hadoop versions.
USAGE
shui versions [--latest]
OPTIONS
--latest Show only the latest available version
GLOBAL OPTIONS
-h (--help) Display this help message
-q (--quiet) Do not output any message
-v (--verbose) Increase the verbosity of messages: "-v" for normal output, "-vv" for more verbose output and "-vvv" for debug
-V (--version) Display this application version
--ansi Force ANSI output
--no-ansi Disable ANSI output
-n (--no-interaction) Do not ask any interactive question
The install
command will download, verify and install a particular Spark/Hadoop version.
USAGE
shui install [--latest] [--spark <...>] [--hadoop <...>] [--target <...>]
OPTIONS
--latest Use the latest available version
--spark Spark version (default: "any")
--hadoop Hadoop version (default: "any")
--target Directory to install into (default: "cwd")
GLOBAL OPTIONS
-h (--help) Display this help message
-q (--quiet) Do not output any message
-v (--verbose) Increase the verbosity of messages: "-v" for normal output, "-vv" for more verbose output and "-vvv" for debug
-V (--version) Display this application version
--ansi Force ANSI output
--no-ansi Disable ANSI output
-n (--no-interaction) Do not ask any interactive question