@@ -33,7 +33,7 @@ alias = spark-shell,pyspark,spark-submit --class org.apache.spark.examples.Spark
[hadoop]
port = 50070:50070,8088:8088
volumes = {container}conf/java/hadoop:/usr/local/hadoop/etc/hadoop
-command = hadoop
+command = hadoop share
alias = hadoop
slave = 2
hook.start = hadoop