source: other-projects/hathitrust/wcsa/extracted-features-solr/trunk/gslis-cluster/SETUP.bash@ 31275

Last change on this file since 31275 was 31275, checked in by davidb, 7 years ago

Changes to allow gc slave nodes to work with local disk versions of packages like Spark, so it can then use the bigger area /hdfsd05/dbbridge/tmp as the area of RDD spill

File size: 3.1 KB
Line 
1
2short_hostname=`hostname -s`
3
4#export SPARK_MASTER_HOST=gchead
5export SPARK_MASTER_HOST=192.168.64.1
6
7export SPARK_MASTER_URL=spark://$SPARK_MASTER_HOST:7077
8export SPARK_SLAVE_HOSTS="gc0 gc1 gc2 gc3 gc4 gc5 gc6 gc7 gc8 gc9"
9
10if [ "$short_hostname" = "nema" ] ; then
11 export ZOOKEEPER_SERVER=localhost:8181
12 export SOLR_NODES="localhost:8983 localhost:8984 localhost:8985 localhost:8986 localhost:8987 localhost:8988 localhost:8989 localhost:8990 localhost:8991 localhost:8992"
13else
14 export ZOOKEEPER_SERVER=gchead:8181
15 export SOLR_NODES="gc0:8983 gc1:8983 gc2:8983 gc3:8983 gc4:8983 gc5:8983 gc6:8983 gc7:8983 gc8:8983 gc9:8983"
16 #export SOLR_NODES="$SOLR_NODES gc0:8984 gc1:8984 gc2:8984 gc3:8984 gc4:8984 gc5:8984 gc6:8984 gc7:8984 gc8:8984 gc9:8984"
17fi
18
19if [ "$short_hostname" != "nema" ] ; then
20 HDFS_HEAD=hdfs://gchead:9000
21fi
22
23if [ "${short_hostname%[0-9]}" = "gc" ] ; then
24 export HTRC_EF_PACKAGE_HOME="/hdfsd05/dbbridge/gslis-cluster"
25else
26 export HTRC_EF_PACKAGE_HOME=`pwd`
27fi
28
29HTRC_EF_NETWORK_HOME=`pwd`
30
31if [ "$short_hostname" = "gsliscluster1" ] ; then
32 echo ""
33 echo "****"
34fi
35
36export JAVA_HOME="$HTRC_EF_NETWORK_HOME/jdk1.8.0"
37export PATH="$JAVA_HOME/bin:$PATH"
38#export _JAVA_OPTIONS="-Xmx512m"
39#export _JAVA_OPTIONS="-Xmx1024m"
40#export _JAVA_OPTIONS="-Xmx2048m"
41export _JAVA_OPTIONS=
42#export _JAVA_OPTIONS="-XX:+HeapDumpOnOutOfMemoryError"
43
44if [ "$short_hostname" = "gsliscluster1" ] ; then
45 echo "* Added in JDK 1.8 into PATH"
46fi
47
48if [ "$short_hostname" != "nema" ] ; then
49 source SETUP/setup-spark.bash
50fi
51
52source SETUP/setup-zookeeper.bash
53source SETUP/setup-solr.bash
54
55
56export PATH="$HTRC_EF_NETWORK_HOME/SCRIPTS:$PATH"
57if [ "$short_hostname" = "gsliscluster1" ] ; then
58 echo "* Added in HTRC EF PACKAGE remote scripting into PATH"
59fi
60
61if [ "$short_hostname" != "nema" ] ; then
62 spark_conf_slaves="$SPARK_HOME/conf/slaves"
63 if [ ! -f "$spark_conf_slaves" ] ; then
64 echo "****"
65 echo "* Populatig $spark_conf_slaves"
66 echo "* With: $SPARK_SLAVE_HOSTS"
67 echo "****"
68 for s in $SPARK_SLAVE_HOSTS ; do
69 echo $s >> "$spark_conf_slaves"
70 done
71 else
72 slaves=`cat "$spark_conf_slaves" | tr '\n' ' '`
73 if [ "$short_hostname" = "gsliscluster1" ] ; then
74 echo "****"
75 echo "* Spark slaves: $slaves"
76 echo "****"
77 fi
78 fi
79fi
80
81
82zookeeper_config_file="$ZOOKEEPER_HOME/conf/zoo.cfg"
83zookeeper_data_dir="$ZOOKEEPER_HOME/data"
84
85if [ ! -f "$zookeeper_config_file" ] ; then
86 echo "****"
87 echo "* Generating $zookeeper_config_file"
88 cat CONF/zoo.cfg.in | sed "s%@zookeeper-data-dir@%$zookeeper_data_dir%g" > "$zookeeper_config_file"
89
90 if [ ! -d "$zookeeper_data_dir" ] ; then
91 echo "* Creating Zookeeper dataDir:"
92 echo "* $zookeeper_data_dir"
93 mkdir "$zookeeper_data_dir"
94 fi
95 echo "****"
96fi
97
98if [ "$short_hostname" = "gsliscluster1" ] ; then
99 echo "****"
100 echo "* Solr nodes: $SOLR_NODES"
101 echo "****"
102else
103 solr_configsets="$SOLR_TOP_LEVEL_HOME/server/solr/configsets"
104 if [ ! -d "$solr_configsets/htrc_configs" ] ; then
105 echo "Untarring htrc_configs.tar.gz in Solr configtests directory"
106 tar xvzf CONF/htrc_configs.tar.gz -C "$solr_configsets"
107 fi
108fi
Note: See TracBrowser for help on using the repository browser.