相关代码
1 | docker run -ti -d --name spark ubuntu:zte-16 bash |
1 | docker run -ti -d --name cloud1 -h cloud1 --add-host cloud1:172.17.0.2 --add-host cloud2:172.17.0.3 --add-host cloud3:172.17.0.4 spark:latest bash |
python update1
2
3
4
5
6
7
8
9
10
11
12
13
14
15tar -xf Python-2.7.13.tgz
rm Python-2.7.13.tgz
echo "export PYTHONHOME=/usr/local/python2" >> /etc/profile
source /etc/profile
apt-get update
apt-get install -y gcc make zlib* libbz2-dev libgdbm-dev liblzma-dev libreadline-dev libsqlite3-dev libssl-dev tcl-dev tk-dev dpkg-dev
pushd Python-2.7.13
./configure --prefix="${PYTHONHOME}" --enable-unicode=ucs4
sed -i 's/#zlib/zlib/g' Modules/Setup
make
make install
popd
rm -rf Python-2.7.12
python get-pip.py
rm get-pip.py
安装SSH
修改允许root登录
配置免密钥登录
1 | 在三台机器上分别执行,每条指令单独执行 |
分别配置hostname
配置host
1 | 分别执行 |
安装Java
1 | cd /usr/local |
安装hadoop
1 | cd /usr/local |
修改配置
hadoop守护进程的运行环境配置
1 | vim ${HADOOP_HOME}/etc/hadoop/hadoop-env.sh |
1 | export JAVA_HOME=/usr/local/jdk1.8.0_131 |
core-site.xml:
1 | mkdir /usr/local/hadoop-2.8.0/tmp |
1 | <configuration> |
hdfs-site.xml:配置namenode和datanode存储命名空间和log的路径
1 | mkdir /usr/local/hadoop-2.8.0/hdfs |
1 | <configuration> |
mapred-site.xml:配置MapReduce。
1 | vim ${HADOOP_HOME}/etc/hadoop/mapred-site.xml |
1 | <configuration> |
yarn-site.xml:配置resourcesmanager和nodemanager
1 | vim ${HADOOP_HOME}/etc/hadoop/yarn-site.xml |
1 | <configuration> |
slaves:配置集群的DataNode节点,这些节点是slaves,NameNode是Master。在conf/slaves文件中列出所有slave的主机名或者IP地址,一行一个。配置如下:
1 | vim ${HADOOP_HOME}/etc/hadoop/slaves |
1 | Slave1 |
启动hadoop
Master
1 | hdfs namenode -format |
启动正确后
1 | Master显示 |
查看存活的端口
1 | http://10.42.10.61:8088/ |
1 | http://10.42.10.61:50070 |
安装Scala Spark
1 | cd /usr/local |
配置修改
spark-env.sh:spark执行任务的环境配置,需要根据自己的机器配置来设置,内存和核心数配置的时候主要不要超出虚拟机的配置,尤其是存在默认值的配置需要仔细查看,修改。
1 | cp ${SPARK_HOME}/conf/spark-env.sh.template ${SPARK_HOME}/conf/spark-env.sh |
1 | export SPARK_DIST_CLASSPATH=$(/usr/local/hadoop-2.8.0/bin/hadoop classpath) |
spark-default.conf:
1 | cp ${SPARK_HOME}/conf/spark-defaults.conf.template ${SPARK_HOME}/conf/spark-defaults.conf |
1 | spark.master spark://Master:7077 |
Slaves
1 | cp ${SPARK_HOME}/conf/slaves.template ${SPARK_HOME}/conf/slaves |
1 | Master |
1 | scp -r ${SPARK_HOME} root@Slave1:/usr/local |
1 | hadoop fs -mkdir -p /spark/history |
安装新版本Pip
1 | apt-get update |
拷贝pip-9.0.1-py2.py3-none-any.whl文件
配置公司pip源
1 | pip install pip-9.0.1-py2.py3-none-any.whl |
tensorflow
1 | pip2 install tensorflow |
Run
1 | rm -rf mnist_model/ |
1 | echo "hello world hello Hello" > test.txt |