ansible-spark
ansible-spark copied to clipboard
An Ansible role for installing Apache Spark.
ansible-spark
An Ansible role for installing Apache Spark.
Role Variables
spark_version- Spark version.spark_install_java- flag toggling the JDK installation using the builtin azavea.java role dependency (default:yes)spark_cloudera_distribution- Cloudera distribution version (default:cdh5.4)spark_symlinks_enabled(defaultyes) - ifyesdeploy 2 symlinks (<spark_home>/conf -> /etc/spark/conf ; <spark_home> ->spark_usr_dir)spark_shims_enabled(defaultyes) - ifyesdeploy the shims (like/usr/bin/spark-shell,/usr/bin/spark-submit)spark_env_extras- An optional dictionary with key and value attributes to add tospark-env.sh(e.g.MESOS_NATIVE_LIBRARY: "/usr/local/lib/libmesos.so")spark_defaults_extras- An optional dictionary with key and value attributes to add tospark-defaults.conf(e.g."spark.eventLog.enabled": true)spark_user_groups- an optional list of (OS)groups the spark user should belong tospark_user_shell- the spark user's default shell (default:/bin/false)spark_log4j_loggers- A list of dictionaries configuring the spark log4j loggers (default: logger config from spark/conf/log4j.properties.template)
Example Playbook
See the examples directory.