rabin 7 yıl önce
ebeveyn
işleme
8ea064ef9c

+ 2 - 1
src/docker/build/java/base/init.sh

@@ -5,4 +5,5 @@ start_sshd()
 	/usr/sbin/sshd
 }
 check jdk lib
-hosts
+hosts
+start_sshd

+ 5 - 5
src/docker/build/java/hadoop/hadoop.sh

@@ -18,11 +18,11 @@ start_hadoop()
     fi
     rm -rf /root/hdfs/*
     rm -rf /root/hadoop/tmp/*
-    $HADOOP_HOME/bin/hdfs namenode -format
-    #$HADOOP_HOME/sbin/start-dfs.sh
-    #$HADOOP_HOME/sbin/start-yarn.sh
-    $HADOOP_HOME/sbin/start-all.sh
-}
+    $HADOOP_HOME/bin/hdfs namenode -format &
+    $HADOOP_HOME/sbin/start-dfs.sh &
+    $HADOOP_HOME/sbin/start-yarn.sh &
+    #$HADOOP_HOME/sbin/start-all.sh &
+} 
 
 stop_hadoop()
 {

+ 1 - 1
src/docker/conf/data.conf

@@ -33,7 +33,7 @@ alias = spark-shell,pyspark,spark-submit --class org.apache.spark.examples.Spark
 [hadoop]
 port = 50070:50070,8088:8088
 volumes = {container}conf/java/hadoop:/usr/local/hadoop/etc/hadoop
-command = hadoop share
+command = hadoop
 alias = hadoop
 slave = 2
 hook.start = hadoop