1
# A Hortonworks HDP 2.1 HIVE, mysql, and Hadoop Cluster
3
This bundle is a 7 node Hadoop cluster designed to scale out. It contains the following units:
5
One Hadoop Master (yarn & hdfs) Node
11
Once you have a cluster running, just run:
12
1) juju ssh yarn-hdfs-master/0 <<= ssh to hadoop master
13
2) Smoke test HDFS admin functionality- As the HDFS user, create a /user/$CLIENT_USER in
14
hadoop file system - Below steps verifies/demos HDFS functionality
16
b) hdfs dfs -mkdir -p /user/ubuntu
17
c) hdfs dfs -chown ubuntu:ubuntu /user/ubuntu
18
d) hdfs dfs -chmod -R 755 /user/ubuntu
21
3) Smoke test YARN and Mapreduce - Run the smoke test as the $CLIENT_USER, using Terasort and sort 10GB of data.
22
a) hadoop jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples-*.jar teragen 10000 /user/ubuntu/teragenout
23
b) hadoop jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples-*.jar terasort /user/ubuntu/teragenout /user/ubuntu/terasortout
25
4) Smoke test HDFS funtionality from ubuntu user space - delete mapreduce output from hdfs
26
hdfs dfs -rm -r /user/ubuntu/teragenout
29
1) juju ssh hdphive/0 <<= ssh to hive server
34
create table test(col1 int, col2 string);
37
5) exit from $HIVE_USER session
39
7) hadoop dfsadmin -report <<== verify connection to the remote HDFS cluster
40
7) hdfs dfs -ls <<== verify that "test" directory has been created on the remote HDFS cluster
43
In order to increase the amount of slaves, you must add units, to add one unit:
44
juju add-unit compute-node
45
Or you can add multiple units at once:
46
juju add-unit -n4 compute-node