zeppelin-env.sh 6.3 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192
  1. #!/bin/bash
  2. #
  3. # Licensed to the Apache Software Foundation (ASF) under one or more
  4. # contributor license agreements. See the NOTICE file distributed with
  5. # this work for additional information regarding copyright ownership.
  6. # The ASF licenses this file to You under the Apache License, Version 2.0
  7. # (the "License"); you may not use this file except in compliance with
  8. # the License. You may obtain a copy of the License at
  9. #
  10. # http://www.apache.org/licenses/LICENSE-2.0
  11. #
  12. # Unless required by applicable law or agreed to in writing, software
  13. # distributed under the License is distributed on an "AS IS" BASIS,
  14. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  15. # See the License for the specific language governing permissions and
  16. # limitations under the License.
  17. #
  18. # export JAVA_HOME=
  19. # export MASTER= # Spark master url. eg. spark://master_addr:7077. Leave empty if you want to use local mode.
  20. # export ZEPPELIN_JAVA_OPTS # Additional jvm options. for example, export ZEPPELIN_JAVA_OPTS="-Dspark.executor.memory=8g -Dspark.cores.max=16"
  21. # export ZEPPELIN_MEM # Zeppelin jvm mem options Default -Xms1024m -Xmx1024m -XX:MaxPermSize=512m
  22. # export ZEPPELIN_INTP_MEM # zeppelin interpreter process jvm mem options. Default -Xms1024m -Xmx1024m -XX:MaxPermSize=512m
  23. # export ZEPPELIN_INTP_JAVA_OPTS # zeppelin interpreter process jvm options.
  24. # export ZEPPELIN_SSL_PORT # ssl port (used when ssl environment variable is set to true)
  25. # export ZEPPELIN_LOG_DIR # Where log files are stored. PWD by default.
  26. # export ZEPPELIN_PID_DIR # The pid files are stored. ${ZEPPELIN_HOME}/run by default.
  27. # export ZEPPELIN_WAR_TEMPDIR # The location of jetty temporary directory.
  28. # export ZEPPELIN_NOTEBOOK_DIR # Where notebook saved
  29. # export ZEPPELIN_NOTEBOOK_HOMESCREEN # Id of notebook to be displayed in homescreen. ex) 2A94M5J1Z
  30. # export ZEPPELIN_NOTEBOOK_HOMESCREEN_HIDE # hide homescreen notebook from list when this value set to "true". default "false"
  31. # export ZEPPELIN_NOTEBOOK_S3_BUCKET # Bucket where notebook saved
  32. # export ZEPPELIN_NOTEBOOK_S3_ENDPOINT # Endpoint of the bucket
  33. # export ZEPPELIN_NOTEBOOK_S3_USER # User in bucket where notebook saved. For example bucket/user/notebook/2A94M5J1Z/note.json
  34. # export ZEPPELIN_NOTEBOOK_S3_KMS_KEY_ID # AWS KMS key ID
  35. # export ZEPPELIN_NOTEBOOK_S3_KMS_KEY_REGION # AWS KMS key region
  36. # export ZEPPELIN_IDENT_STRING # A string representing this instance of zeppelin. $USER by default.
  37. # export ZEPPELIN_NICENESS # The scheduling priority for daemons. Defaults to 0.
  38. # export ZEPPELIN_INTERPRETER_LOCALREPO # Local repository for interpreter's additional dependency loading
  39. # export ZEPPELIN_INTERPRETER_DEP_MVNREPO # Remote principal repository for interpreter's additional dependency loading
  40. # export ZEPPELIN_HELIUM_NPM_REGISTRY # Remote Npm registry for Helium dependency loader
  41. # export ZEPPELIN_NOTEBOOK_STORAGE # Refers to pluggable notebook storage class, can have two classes simultaneously with a sync between them (e.g. local and remote).
  42. # export ZEPPELIN_NOTEBOOK_ONE_WAY_SYNC # If there are multiple notebook storages, should we treat the first one as the only source of truth?
  43. # export ZEPPELIN_NOTEBOOK_PUBLIC # Make notebook public by default when created, private otherwise
  44. #### Spark interpreter configuration ####
  45. ## Use provided spark installation ##
  46. ## defining SPARK_HOME makes Zeppelin run spark interpreter process using spark-submit
  47. ##
  48. # export SPARK_HOME # (required) When it is defined, load it instead of Zeppelin embedded Spark libraries
  49. # export SPARK_SUBMIT_OPTIONS # (optional) extra options to pass to spark submit. eg) "--driver-memory 512M --executor-memory 1G".
  50. # export SPARK_APP_NAME # (optional) The name of spark application.
  51. ## Use embedded spark binaries ##
  52. ## without SPARK_HOME defined, Zeppelin still able to run spark interpreter process using embedded spark binaries.
  53. ## however, it is not encouraged when you can define SPARK_HOME
  54. ##
  55. # Options read in YARN client mode
  56. # export HADOOP_CONF_DIR # yarn-site.xml is located in configuration directory in HADOOP_CONF_DIR.
  57. # Pyspark (supported with Spark 1.2.1 and above)
  58. # To configure pyspark, you need to set spark distribution's path to 'spark.home' property in Interpreter setting screen in Zeppelin GUI
  59. # export PYSPARK_PYTHON # path to the python command. must be the same path on the driver(Zeppelin) and all workers.
  60. # export PYTHONPATH
  61. ## Spark interpreter options ##
  62. ##
  63. # export ZEPPELIN_SPARK_USEHIVECONTEXT # Use HiveContext instead of SQLContext if set true. true by default.
  64. # export ZEPPELIN_SPARK_CONCURRENTSQL # Execute multiple SQL concurrently if set true. false by default.
  65. # export ZEPPELIN_SPARK_IMPORTIMPLICIT # Import implicits, UDF collection, and sql if set true. true by default.
  66. # export ZEPPELIN_SPARK_MAXRESULT # Max number of Spark SQL result to display. 1000 by default.
  67. # export ZEPPELIN_WEBSOCKET_MAX_TEXT_MESSAGE_SIZE # Size in characters of the maximum text message to be received by websocket. Defaults to 1024000
  68. #### HBase interpreter configuration ####
  69. ## To connect to HBase running on a cluster, either HBASE_HOME or HBASE_CONF_DIR must be set
  70. # export HBASE_HOME= # (require) Under which HBase scripts and configuration should be
  71. # export HBASE_CONF_DIR= # (optional) Alternatively, configuration directory can be set to point to the directory that has hbase-site.xml
  72. #### ZeppelinHub connection configuration ####
  73. # export ZEPPELINHUB_API_ADDRESS # Refers to the address of the ZeppelinHub service in use
  74. # export ZEPPELINHUB_API_TOKEN # Refers to the Zeppelin instance token of the user
  75. # export ZEPPELINHUB_USER_KEY # Optional, when using Zeppelin with authentication.
  76. #### Zeppelin impersonation configuration
  77. # export ZEPPELIN_IMPERSONATE_CMD # Optional, when user want to run interpreter as end web user. eg) 'sudo -H -u ${ZEPPELIN_IMPERSONATE_USER} bash -c '
  78. # export ZEPPELIN_IMPERSONATE_SPARK_PROXY_USER #Optional, by default is true; can be set to false if you don't want to use --proxy-user option with Spark interpreter when impersonation enabled
  79. export MASTER=yarn-client
  80. #export SPARK_SUBMIT_OPTIONS="--packages com.databricks:spark-csv_2.10:1.2.0"