forked from PipelineAI/pipeline
-
Notifications
You must be signed in to change notification settings - Fork 0
/
flux-start.sh
executable file
·66 lines (46 loc) · 2.21 KB
/
flux-start.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
#!/bin/bash
echo ...Starting ElasticSearch...
nohup elasticsearch -p $ELASTICSEARCH_HOME/RUNNING_PID &
echo ...Starting Logstash...
nohup logstash agent -f $PIPELINE_HOME/config/logstash/logstash.conf &
echo ...Starting SSH...
service ssh start
echo ...Starting Ganglia...
service ganglia-monitor start
service gmetad start
echo ...Starting Apache2 Httpd...
service apache2 start
echo ...Starting MySQL...
service mysql start
echo ...Starting Cassandra...
nohup cassandra
echo ...Starting ZooKeeper...
nohup zookeeper-server-start $PIPELINE_HOME/config/kafka/zookeeper.properties &
echo ...Starting Redis...
nohup redis-server &
echo ...Starting Tachyon...
nohup $TACHYON_HOME/bin/tachyon-start.sh local
echo ...Starting Kafka...
nohup kafka-server-start $PIPELINE_HOME/config/kafka/server.properties &
echo ...Starting Zeppelin...
nohup $ZEPPELIN_HOME/bin/zeppelin-daemon.sh start
echo ...Starting Spark Master...
nohup $SPARK_HOME/sbin/start-master.sh --webui-port 6060 -i 127.0.0.1 -h 127.0.0.1
echo ...Starting Spark Worker...
nohup $SPARK_HOME/sbin/start-slave.sh --webui-port 6061 spark:https://127.0.0.1:7077
echo ...Starting Apache Spark JDBC ODBC Hive ThriftServer...
## MySql must be started - and the password set - before ThriftServer will startup
## Starting the ThriftServer will create a dummy derby.log and metastore_db per https://github.com/apache/spark/pull/6314
## The actual Hive metastore defined in conf/hive-site.xml is still used, however.
nohup $SPARK_HOME/sbin/start-thriftserver.sh --jars $MYSQL_CONNECTOR_JAR --master spark:https://127.0.0.1:7077
echo ...Starting Spark Notebook...
screen -m -d -S "snb" bash -c 'source ~/pipeline/config/bash/.profile && spark-notebook -Dconfig.file=$PIPELINE_HOME/config/spark-notebook/application-pipeline.conf >> nohup.out'
echo ...Starting Spark History Server...
$SPARK_HOME/sbin/start-history-server.sh
echo ...Starting Kibana...
nohup kibana &
echo ...Starting Kafka Schema Registry...
# Starting this at the end due to race conditions with other kafka components
nohup schema-registry-start $PIPELINE_HOME/config/schema-registry/schema-registry.properties &
echo ...Starting Kafka REST Proxy...
nohup kafka-rest-start $PIPELINE_HOME/config/kafka-rest/kafka-rest.properties &