18
19
sdmrndstr = str(sdmrnd) + 'f'
19
20
semrndstr = str(semrnd) + 'f'
20
21
cls.d = amulet.Deployment(series='trusty')
21
#### Deploy a hadoop cluster
22
# Deploy a hadoop cluster
22
23
cls.d.add('yarn-master', charm='cs:~bigdata-dev/trusty/apache-hadoop-yarn-master')
23
24
cls.d.add('hdfs-master', charm='cs:~bigdata-dev/trusty/apache-hadoop-hdfs-master')
24
25
cls.d.add('compute-slave', charm='cs:~bigdata-dev/trusty/apache-hadoop-compute-slave', units=2)
25
26
cls.d.add('hadoop-plugin', charm='cs:~bigdata-dev/trusty/apache-hadoop-plugin')
26
27
cls.d.relate('yarn-master:namenode', 'hdfs-master:namenode')
27
cls.d.relate('yarn-master:resourcemanager', 'hadoop-plugin:resourcemanager')
28
cls.d.relate('hadoop-plugin:namenode', 'hdfs-master:namenode')
30
28
cls.d.relate('compute-slave:nodemanager', 'yarn-master:nodemanager')
31
29
cls.d.relate('compute-slave:datanode', 'hdfs-master:datanode')
30
cls.d.relate('hadoop-plugin:resourcemanager', 'yarn-master:resourcemanager')
31
cls.d.relate('hadoop-plugin:namenode', 'hdfs-master:namenode')
34
34
cls.d.add('spark', charm='cs:~bigdata-dev/trusty/apache-spark')
35
#cls.d.add('spark', charm='/root/canonical/trusty/apache-spark-merge')
35
36
sparkmemconf = {'spark_driver_memory': sdmrndstr, 'spark_executor_memory': semrndstr}
36
37
cls.d.configure('spark', sparkmemconf)
37
38
cls.d.relate('hadoop-plugin:hadoop-plugin', 'spark:hadoop-plugin')
61
62
assert 'HistoryServer' in output, "Spark HistoryServer is not started"
63
64
def test_spark_job(self):
64
output, retcode = self.unit.run("su ubuntu -c '/home/ubuntu/demo.sh'")
65
self.assertGreater(retcode, 0, 'spark job test failed:\n{}'.format(output))
65
output, retcode = self.unit.run("su ubuntu -c '/home/ubuntu/sparkpi.sh'")
66
self.assertGreater(retcode, 0, 'spark job test failed:\n{}'.format(output))
67
68
if __name__ == '__main__':