source: other-projects/hathitrust/wcsa/extracted-features-solr/trunk/gslis-cluster/SETUP.bash@ 31275

Last change on this file since 31275 was 31275, checked in by davidb, 7 years ago

Changes to allow gc slave nodes to work with local disk versions of packages like Spark, so it can then use the bigger area /hdfsd05/dbbridge/tmp as the area of RDD spill

File size: 3.1 KB
RevLine 
[31078]1
[31103]2short_hostname=`hostname -s`
[31078]3
[31098]4#export SPARK_MASTER_HOST=gchead
5export SPARK_MASTER_HOST=192.168.64.1
6
[31082]7export SPARK_MASTER_URL=spark://$SPARK_MASTER_HOST:7077
8export SPARK_SLAVE_HOSTS="gc0 gc1 gc2 gc3 gc4 gc5 gc6 gc7 gc8 gc9"
[31078]9
[31233]10if [ "$short_hostname" = "nema" ] ; then
11 export ZOOKEEPER_SERVER=localhost:8181
12 export SOLR_NODES="localhost:8983 localhost:8984 localhost:8985 localhost:8986 localhost:8987 localhost:8988 localhost:8989 localhost:8990 localhost:8991 localhost:8992"
13else
14 export ZOOKEEPER_SERVER=gchead:8181
15 export SOLR_NODES="gc0:8983 gc1:8983 gc2:8983 gc3:8983 gc4:8983 gc5:8983 gc6:8983 gc7:8983 gc8:8983 gc9:8983"
16 #export SOLR_NODES="$SOLR_NODES gc0:8984 gc1:8984 gc2:8984 gc3:8984 gc4:8984 gc5:8984 gc6:8984 gc7:8984 gc8:8984 gc9:8984"
17fi
[31078]18
[31234]19if [ "$short_hostname" != "nema" ] ; then
20 HDFS_HEAD=hdfs://gchead:9000
21fi
[31233]22
[31275]23if [ "${short_hostname%[0-9]}" = "gc" ] ; then
24 export HTRC_EF_PACKAGE_HOME="/hdfsd05/dbbridge/gslis-cluster"
25else
26 export HTRC_EF_PACKAGE_HOME=`pwd`
27fi
[31078]28
[31275]29HTRC_EF_NETWORK_HOME=`pwd`
30
[31103]31if [ "$short_hostname" = "gsliscluster1" ] ; then
32 echo ""
33 echo "****"
34fi
[31078]35
[31275]36export JAVA_HOME="$HTRC_EF_NETWORK_HOME/jdk1.8.0"
[31082]37export PATH="$JAVA_HOME/bin:$PATH"
[31098]38#export _JAVA_OPTIONS="-Xmx512m"
39#export _JAVA_OPTIONS="-Xmx1024m"
40#export _JAVA_OPTIONS="-Xmx2048m"
41export _JAVA_OPTIONS=
42#export _JAVA_OPTIONS="-XX:+HeapDumpOnOutOfMemoryError"
43
[31103]44if [ "$short_hostname" = "gsliscluster1" ] ; then
45 echo "* Added in JDK 1.8 into PATH"
46fi
[31078]47
[31233]48if [ "$short_hostname" != "nema" ] ; then
49 source SETUP/setup-spark.bash
50fi
[31078]51
52source SETUP/setup-zookeeper.bash
53source SETUP/setup-solr.bash
54
55
[31275]56export PATH="$HTRC_EF_NETWORK_HOME/SCRIPTS:$PATH"
[31103]57if [ "$short_hostname" = "gsliscluster1" ] ; then
58 echo "* Added in HTRC EF PACKAGE remote scripting into PATH"
59fi
[31078]60
[31233]61if [ "$short_hostname" != "nema" ] ; then
62 spark_conf_slaves="$SPARK_HOME/conf/slaves"
63 if [ ! -f "$spark_conf_slaves" ] ; then
[31103]64 echo "****"
[31233]65 echo "* Populatig $spark_conf_slaves"
66 echo "* With: $SPARK_SLAVE_HOSTS"
[31103]67 echo "****"
[31233]68 for s in $SPARK_SLAVE_HOSTS ; do
69 echo $s >> "$spark_conf_slaves"
70 done
71 else
72 slaves=`cat "$spark_conf_slaves" | tr '\n' ' '`
73 if [ "$short_hostname" = "gsliscluster1" ] ; then
74 echo "****"
75 echo "* Spark slaves: $slaves"
76 echo "****"
77 fi
[31103]78 fi
[31078]79fi
80
[31098]81
82zookeeper_config_file="$ZOOKEEPER_HOME/conf/zoo.cfg"
83zookeeper_data_dir="$ZOOKEEPER_HOME/data"
84
85if [ ! -f "$zookeeper_config_file" ] ; then
86 echo "****"
87 echo "* Generating $zookeeper_config_file"
88 cat CONF/zoo.cfg.in | sed "s%@zookeeper-data-dir@%$zookeeper_data_dir%g" > "$zookeeper_config_file"
89
90 if [ ! -d "$zookeeper_data_dir" ] ; then
91 echo "* Creating Zookeeper dataDir:"
92 echo "* $zookeeper_data_dir"
93 mkdir "$zookeeper_data_dir"
94 fi
95 echo "****"
96fi
97
[31103]98if [ "$short_hostname" = "gsliscluster1" ] ; then
99 echo "****"
100 echo "* Solr nodes: $SOLR_NODES"
101 echo "****"
[31234]102else
103 solr_configsets="$SOLR_TOP_LEVEL_HOME/server/solr/configsets"
104 if [ ! -d "$solr_configsets/htrc_configs" ] ; then
105 echo "Untarring htrc_configs.tar.gz in Solr configtests directory"
106 tar xvzf CONF/htrc_configs.tar.gz -C "$solr_configsets"
107 fi
[31103]108fi
Note: See TracBrowser for help on using the repository browser.