|
@@ -21,7 +21,7 @@
|
|
|
# A comma separated list of machine hostname or IP would be installed DolphinScheduler,
|
|
|
# including master, worker, api, alert. If you want to deploy in pseudo-distributed
|
|
|
# mode, just write a pseudo-distributed hostname
|
|
|
-# Example for hostnames: ips="ds1,ds2,ds3,ds4,ds5", Example for IP: ips="192.168.8.1,192.168.8.2,192.168.8.3,192.168.8.4,192.168.8.5"
|
|
|
+# Example for hostnames: ips="ds1,ds2,ds3,ds4,ds5", Example for IPs: ips="192.168.8.1,192.168.8.2,192.168.8.3,192.168.8.4,192.168.8.5"
|
|
|
ips="ds1,ds2,ds3,ds4,ds5"
|
|
|
|
|
|
# Port of SSH protocol, default value is 22. For now we only support same port in all `ips` machine
|
|
@@ -30,29 +30,29 @@ sshPort="22"
|
|
|
|
|
|
# A comma separated list of machine hostname or IP would be installed Master server, it
|
|
|
# must be a subset of configuration `ips`.
|
|
|
-# Example for hostnames: ips="ds1,ds2", Example for IP: ips="192.168.8.1,192.168.8.2"
|
|
|
+# Example for hostnames: masters="ds1,ds2", Example for IPs: masters="192.168.8.1,192.168.8.2"
|
|
|
masters="ds1,ds2"
|
|
|
|
|
|
# A comma separated list of machine <hostname>:<workerGroup> or <IP>:<workerGroup>.All hostname or IP must be a
|
|
|
# subset of configuration `ips`, And workerGroup have default value as `default`, but we recommend you declare behind the hosts
|
|
|
-# Example for hostnames: ips="ds1:default,ds2:default,ds3:default", Example for IP: ips="192.168.8.1:default,192.168.8.2:default,192.168.8.3:default"
|
|
|
+# Example for hostnames: workers="ds1:default,ds2:default,ds3:default", Example for IPs: workers="192.168.8.1:default,192.168.8.2:default,192.168.8.3:default"
|
|
|
workers="ds1:default,ds2:default,ds3:default,ds4:default,ds5:default"
|
|
|
|
|
|
# A comma separated list of machine hostname or IP would be installed Alert server, it
|
|
|
# must be a subset of configuration `ips`.
|
|
|
-# Example for hostnames: ips="ds3", Example for IP: ips="192.168.8.3"
|
|
|
+# Example for hostname: alertServer="ds3", Example for IP: alertServer="192.168.8.3"
|
|
|
alertServer="ds3"
|
|
|
|
|
|
# A comma separated list of machine hostname or IP would be installed API server, it
|
|
|
# must be a subset of configuration `ips`.
|
|
|
-# Example for hostnames: ips="ds1", Example for IP: ips="192.168.8.1"
|
|
|
+# Example for hostname: apiServers="ds1", Example for IP: apiServers="192.168.8.1"
|
|
|
apiServers="ds1"
|
|
|
|
|
|
-# The directory to install DolphinScheduler for all machine we config above. It will automatically created by `install.sh` script if not exists.
|
|
|
-# **DO NOT** set this configuration same as the current path (pwd)
|
|
|
+# The directory to install DolphinScheduler for all machine we config above. It will automatically be created by `install.sh` script if not exists.
|
|
|
+# Do not set this configuration same as the current path (pwd)
|
|
|
installPath="/data1_1T/dolphinscheduler"
|
|
|
|
|
|
-# The user to deploy DolphinScheduler for all machine we config above. For now user must create by yourself before run `install.sh`
|
|
|
+# The user to deploy DolphinScheduler for all machine we config above. For now user must create by yourself before running `install.sh`
|
|
|
# script. The user needs to have sudo privileges and permissions to operate hdfs. If hdfs is enabled than the root directory needs
|
|
|
# to be created by this user
|
|
|
deployUser="dolphinscheduler"
|
|
@@ -67,7 +67,7 @@ dataBasedirPath="/tmp/dolphinscheduler"
|
|
|
# and this configuration only support one parameter so far.
|
|
|
javaHome="/your/java/home/here"
|
|
|
|
|
|
-# DolphinScheduler API service port, also this your DolphinScheduler UI component's URL port, default values is 12345
|
|
|
+# DolphinScheduler API service port, also this is your DolphinScheduler UI component's URL port, default value is 12345
|
|
|
apiServerPort="12345"
|
|
|
|
|
|
# ---------------------------------------------------------
|
|
@@ -122,12 +122,12 @@ taskPluginDir="lib/plugin/task"
|
|
|
# resource storage type: HDFS, S3, NONE
|
|
|
resourceStorageType="NONE"
|
|
|
|
|
|
-# resource store on HDFS/S3 path, resource file will store to this hadoop hdfs path, self configuration, please make sure the directory exists on hdfs and have read write permissions. "/dolphinscheduler" is recommended
|
|
|
+# resource store on HDFS/S3 path, resource file will store to this hdfs path, self configuration, please make sure the directory exists on hdfs and has read write permissions. "/dolphinscheduler" is recommended
|
|
|
resourceUploadPath="/dolphinscheduler"
|
|
|
|
|
|
-# if resourceStorageType is HDFS,defaultFS write namenode address,HA you need to put core-site.xml and hdfs-site.xml in the conf directory.
|
|
|
+# if resourceStorageType is HDFS,defaultFS write namenode address,HA, you need to put core-site.xml and hdfs-site.xml in the conf directory.
|
|
|
# if S3,write S3 address,HA,for example :s3a://dolphinscheduler,
|
|
|
-# Note,s3 be sure to create the root directory /dolphinscheduler
|
|
|
+# Note,S3 be sure to create the root directory /dolphinscheduler
|
|
|
defaultFS="hdfs://mycluster:8020"
|
|
|
|
|
|
# if resourceStorageType is S3, the following three configuration is required, otherwise please ignore
|
|
@@ -138,13 +138,13 @@ s3SecretKey="xxxxxxxxxx"
|
|
|
# resourcemanager port, the default value is 8088 if not specified
|
|
|
resourceManagerHttpAddressPort="8088"
|
|
|
|
|
|
-# if resourcemanager HA is enabled, please set the HA IPs; if resourcemanager is single, keep this value empty
|
|
|
+# if resourcemanager HA is enabled, please set the HA IPs; if resourcemanager is single node, keep this value empty
|
|
|
yarnHaIps="192.168.xx.xx,192.168.xx.xx"
|
|
|
|
|
|
-# if resourcemanager HA is enabled or not use resourcemanager, please keep the default value; If resourcemanager is single, you only need to replace ds1 to actual resourcemanager hostname
|
|
|
+# if resourcemanager HA is enabled or not use resourcemanager, please keep the default value; If resourcemanager is single node, you only need to replace 'yarnIp1' to actual resourcemanager hostname
|
|
|
singleYarnIp="yarnIp1"
|
|
|
|
|
|
-# who have permissions to create directory under HDFS/S3 root path
|
|
|
+# who has permission to create directory under HDFS/S3 root path
|
|
|
# Note: if kerberos is enabled, please config hdfsRootUser=
|
|
|
hdfsRootUser="hdfs"
|
|
|
|