December
16th,
2017
install-spark is a fully automated Spark installation bash script for linux Dabian based Servers.
Script will install following key components needes to install Real-time data processing with few clicks.
- Full Update & Upgrade
- install up-to-date JRE
- install up-to-date JDK
- install scala 2.11.11
- install Apache kafka 2.10 0.10.2.1
- install git
- install sbt
- install Apache spark spark-2.2.0 pre-built for Apache hadoop 2.7 & later
- all the environmental + all the PATH variables
Using the Script
1. Granting the executing permission to the script
Navigate to the directory where the script is.
$ chmod +x install-spark.sh
2. Run the script
$ source install-spark.sh
Customizing the Script
In the script you can define the versions of libraries that you need to install.
eg: spark-2.2.0-bin-hadoop2.7 , scala-2.11.11
Under the # Default Application versions
define your prefered version.
# Default Application versions
SCALA_VERSION="scala-2.11.11"
SPARK_VERSION="spark-2.2.0-bin-hadoop2.7"
KAFKA_VERSION="kafka_2.10-0.10.2.1"
- apache (3) ,
- kafka (3) ,
- bigdata (3) ,
- scala (2) ,
- Real-Time-Processing (2) ,
- Spark (2) ,
- streaming (2) ,
- java (1) ,
- git (2)