forked from kiwenlau/hadoop-cluster-docker
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathdocker-stack.yml
74 lines (73 loc) · 3.5 KB
/
docker-stack.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
version: "3.2"
services:
hadoop-master:
image: xiaoqiaotq/hadoop:0.2
hostname: hadoop-master
ports:
- { published: 9000, target: 9000, mode: host } # hdfs
- { published: 50070, target: 50070, mode: host } # namenode ui
- { published: 50075, target: 50075, mode: host } #datanode http ui
- { published: 50010, target: 50010, mode: host } #datanode address
- { published: 50020, target: 50020, mode: host } #datanode ipc.address
- { published: 8088, target: 8088, mode: host } # resourcemanager ui
- { published: 8030, target: 8030, mode: host } # resourcemanager
- { published: 8031, target: 8031, mode: host } # resourcemanager
- { published: 8032, target: 8032, mode: host } # resourcemanager
- { published: 8033, target: 8033, mode: host } # resourcemanager
- { published: 8042, target: 8042, mode: host } #nodemanager ui
- { published: 16010, target: 16010, mode: host }
- { published: 10002, target: 10002, mode: host } #hiveserver2 ui
- { published: 9083, target: 9083, mode: host } # hive meta store
- { published: 19888, target: 19888, mode: host } # map reduce history server
- { published: 7077, target: 7077, mode: host } # spark master
- { published: 8080, target: 8080, mode: host } # spark master ui
- { published: 8081, target: 8081, mode: host } # spark worker ui
- { published: 4040, target: 4040, mode: host } # spark job detail
- { published: 18080, target: 18080, mode: host } # spark history job
- { published: 8998, target: 8998, mode: host } # spark livy
deploy:
placement:
constraints: [node.role == manager]
hadoop-slave:
image: xiaoqiaotq/hadoop:0.2
ports:
- { published: 9000, target: 9000, mode: host } # hdfs
- { published: 50070, target: 50070, mode: host } # namenode ui
- { published: 50075, target: 50075, mode: host } #datanode http ui
- { published: 50010, target: 50010, mode: host } #datanode address
- { published: 50020, target: 50020, mode: host } #datanode ipc.address
- { published: 8088, target: 8088, mode: host } # resourcemanager ui
- { published: 8042, target: 8042, mode: host } #nodemanager ui
- { published: 16010, target: 16010, mode: host }
- { published: 10002, target: 10002, mode: host } #hiveserver2 ui
- { published: 9083, target: 9083, mode: host } # hive meta store
- { published: 19888, target: 19888, mode: host } # map reduce history server
- { published: 7077, target: 7077, mode: host } # spark master
- { published: 8080, target: 8080, mode: host } # spark master ui
- { published: 8081, target: 8081, mode: host } # spark worker ui
- { published: 4040, target: 4040, mode: host } # spark job detail
- { published: 18080, target: 18080, mode: host } # spark history job
- { published: 8998, target: 8998, mode: host } # spark livy
deploy:
placement:
constraints: [node.role != manager]
mysql:
image: mysql:5.7
environment:
MYSQL_DATABASE: 'hive'
# So you don't have to use root, but you can if you like
MYSQL_USER: 'hive'
# You can use whatever password you like
MYSQL_PASSWORD: 'hive'
# Password for root access
MYSQL_ROOT_PASSWORD: 'admin'
ports:
# <Port exposed> : < MySQL Port running inside container>
- '3306:3306'
volumes:
- mysql:/var/lib/mysql
deploy:
placement:
constraints: [node.role == manager]
volumes:
mysql: