15
15
name : Spark Images
16
16
runs-on : ubuntu-latest
17
17
env :
18
- HADOOP_VERSION : " 2.7"
19
- SCALA_VERSION : " 2.12.11"
20
18
DOCKERHUB_USR : ${{ secrets.DOCKERHUB_USR }}
21
19
DOCKERHUB_PWD : ${{ secrets.DOCKERHUB_PWD }}
22
20
25
23
- name : Checkout Code
26
24
uses : actions/checkout@master
27
25
28
- - name : Build Base Image
26
+ - name : Build Base Image - Scala v2.12.10
27
+ env :
28
+ SCALA_VERSION : " 2.12.10"
29
29
run : |
30
30
cd ${GITHUB_WORKSPACE}/build
31
31
docker build \
@@ -37,14 +37,15 @@ jobs:
37
37
- name : Build Spark Base Image - Spark v3.0.0 (latest)
38
38
env :
39
39
SPARK_VERSION : " 3.0.0"
40
+ HADOOP_VERSION : " 3.2"
40
41
run : |
41
42
cd ${GITHUB_WORKSPACE}/build
42
43
docker build \
43
44
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
44
45
--build-arg spark_version="${SPARK_VERSION}" \
45
46
--build-arg hadoop_version="${HADOOP_VERSION}" \
46
47
-f docker/spark-base/Dockerfile \
47
- -t spark-base:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} .
48
+ -t spark-base:${SPARK_VERSION} .
48
49
49
50
- name : Build & Push Spark Master Image - Spark v3.0.0 (latest)
50
51
env :
@@ -54,13 +55,12 @@ jobs:
54
55
docker build \
55
56
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
56
57
--build-arg spark_version="${SPARK_VERSION}" \
57
- --build-arg hadoop_version="${HADOOP_VERSION}" \
58
58
-f docker/spark-master/Dockerfile \
59
- -t spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} .
59
+ -t spark-master:${SPARK_VERSION} .
60
60
echo ${DOCKERHUB_PWD} | docker login -u ${DOCKERHUB_USR} --password-stdin
61
- docker tag spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION }
62
- docker push ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION}
63
- docker tag spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} ${DOCKERHUB_USR}/spark-master:latest
61
+ docker tag spark-master:${SPARK_VERSION} ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}
62
+ docker push ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}
63
+ docker tag spark-master:${SPARK_VERSION} ${DOCKERHUB_USR}/spark-master:latest
64
64
docker push ${DOCKERHUB_USR}/spark-master:latest
65
65
66
66
- name : Build & Push Spark Worker Image - Spark v3.0.0 (latest)
@@ -71,26 +71,37 @@ jobs:
71
71
docker build \
72
72
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
73
73
--build-arg spark_version="${SPARK_VERSION}" \
74
- --build-arg hadoop_version="${HADOOP_VERSION}" \
75
74
-f docker/spark-worker/Dockerfile \
76
- -t spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} .
75
+ -t spark-worker:${SPARK_VERSION} .
77
76
echo ${DOCKERHUB_PWD} | docker login -u ${DOCKERHUB_USR} --password-stdin
78
- docker tag spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION }
79
- docker push ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION}
80
- docker tag spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} ${DOCKERHUB_USR}/spark-worker:latest
77
+ docker tag spark-worker:${SPARK_VERSION} ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}
78
+ docker push ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}
79
+ docker tag spark-worker:${SPARK_VERSION} ${DOCKERHUB_USR}/spark-worker:latest
81
80
docker push ${DOCKERHUB_USR}/spark-worker:latest
82
81
82
+ - name : Build Base Image - Scala v2.11.12
83
+ env :
84
+ SCALA_VERSION : " 2.11.12"
85
+ run : |
86
+ cd ${GITHUB_WORKSPACE}/build
87
+ docker build \
88
+ --build-arg build_date="$(date -u +'%Y-%m-%d')" \
89
+ --build-arg scala_version="${SCALA_VERSION}" \
90
+ -f docker/base/Dockerfile \
91
+ -t base:latest .
92
+
83
93
- name : Build Spark Base Image - Spark v2.4.4
84
94
env :
85
95
SPARK_VERSION : " 2.4.4"
96
+ HADOOP_VERSION : " 2.7"
86
97
run : |
87
98
cd ${GITHUB_WORKSPACE}/build
88
99
docker build \
89
100
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
90
101
--build-arg spark_version="${SPARK_VERSION}" \
91
102
--build-arg hadoop_version="${HADOOP_VERSION}" \
92
103
-f docker/spark-base/Dockerfile \
93
- -t spark-base:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} .
104
+ -t spark-base:${SPARK_VERSION} .
94
105
95
106
- name : Build & Push Spark Master Image - Spark v2.4.4
96
107
env :
@@ -100,12 +111,11 @@ jobs:
100
111
docker build \
101
112
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
102
113
--build-arg spark_version="${SPARK_VERSION}" \
103
- --build-arg hadoop_version="${HADOOP_VERSION}" \
104
114
-f docker/spark-master/Dockerfile \
105
- -t spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} .
115
+ -t spark-master:${SPARK_VERSION} .
106
116
echo ${DOCKERHUB_PWD} | docker login -u ${DOCKERHUB_USR} --password-stdin
107
- docker tag spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION }
108
- docker push ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION}
117
+ docker tag spark-master:${SPARK_VERSION} ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}
118
+ docker push ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}
109
119
110
120
- name : Build & Push Spark Worker Image - Spark v2.4.4
111
121
env :
@@ -115,24 +125,24 @@ jobs:
115
125
docker build \
116
126
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
117
127
--build-arg spark_version="${SPARK_VERSION}" \
118
- --build-arg hadoop_version="${HADOOP_VERSION}" \
119
128
-f docker/spark-worker/Dockerfile \
120
- -t spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} .
129
+ -t spark-worker:${SPARK_VERSION} .
121
130
echo ${DOCKERHUB_PWD} | docker login -u ${DOCKERHUB_USR} --password-stdin
122
- docker tag spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION }
123
- docker push ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION}
131
+ docker tag spark-worker:${SPARK_VERSION} ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}
132
+ docker push ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}
124
133
125
134
- name : Build Spark Base Image - Spark v2.4.0
126
135
env :
127
136
SPARK_VERSION : " 2.4.0"
137
+ HADOOP_VERSION : " 2.7"
128
138
run : |
129
139
cd ${GITHUB_WORKSPACE}/build
130
140
docker build \
131
141
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
132
142
--build-arg spark_version="${SPARK_VERSION}" \
133
143
--build-arg hadoop_version="${HADOOP_VERSION}" \
134
144
-f docker/spark-base/Dockerfile \
135
- -t spark-base:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} .
145
+ -t spark-base:${SPARK_VERSION} .
136
146
137
147
- name : Build & Push Spark Master Image - Spark v2.4.0
138
148
env :
@@ -142,12 +152,11 @@ jobs:
142
152
docker build \
143
153
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
144
154
--build-arg spark_version="${SPARK_VERSION}" \
145
- --build-arg hadoop_version="${HADOOP_VERSION}" \
146
155
-f docker/spark-master/Dockerfile \
147
- -t spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} .
156
+ -t spark-master:${SPARK_VERSION} .
148
157
echo ${DOCKERHUB_PWD} | docker login -u ${DOCKERHUB_USR} --password-stdin
149
- docker tag spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION }
150
- docker push ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}-hadoop-${HADOOP_VERSION}
158
+ docker tag spark-master:${SPARK_VERSION} ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}
159
+ docker push ${DOCKERHUB_USR}/spark-master:${SPARK_VERSION}
151
160
152
161
- name : Build & Push Spark Worker Image - Spark v2.4.0
153
162
env :
@@ -157,19 +166,17 @@ jobs:
157
166
docker build \
158
167
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
159
168
--build-arg spark_version="${SPARK_VERSION}" \
160
- --build-arg hadoop_version="${HADOOP_VERSION}" \
161
169
-f docker/spark-worker/Dockerfile \
162
- -t spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} .
170
+ -t spark-worker:${SPARK_VERSION} .
163
171
echo ${DOCKERHUB_PWD} | docker login -u ${DOCKERHUB_USR} --password-stdin
164
- docker tag spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION} ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION }
165
- docker push ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}-hadoop-${HADOOP_VERSION}
172
+ docker tag spark-worker:${SPARK_VERSION} ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}
173
+ docker push ${DOCKERHUB_USR}/spark-worker:${SPARK_VERSION}
166
174
167
175
jupyterlab-images :
168
176
169
177
name : JupyterLab Images
170
178
runs-on : ubuntu-latest
171
179
env :
172
- SCALA_VERSION : " 2.12.11"
173
180
JUPYTERLAB_VERSION : " 2.1.4"
174
181
DOCKERHUB_USR : ${{ secrets.DOCKERHUB_USR }}
175
182
DOCKERHUB_PWD : ${{ secrets.DOCKERHUB_PWD }}
@@ -179,7 +186,9 @@ jobs:
179
186
- name : Checkout Code
180
187
uses : actions/checkout@master
181
188
182
- - name : Build Base Image
189
+ - name : Build Base Image - Scala v2.12.10
190
+ env :
191
+ SCALA_VERSION : " 2.12.10"
183
192
run : |
184
193
cd ${GITHUB_WORKSPACE}/build
185
194
docker build \
@@ -191,13 +200,16 @@ jobs:
191
200
- name : Build & Push JupyterLab Image - Spark v3.0.0 (latest)
192
201
env :
193
202
SPARK_VERSION : " 3.0.0"
203
+ SCALA_VERSION : " 2.12.10"
204
+ SCALA_KERNEL_VERSION : " 0.10.9"
194
205
run : |
195
206
cd ${GITHUB_WORKSPACE}/build
196
207
docker build \
197
208
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
198
209
--build-arg scala_version="${SCALA_VERSION}" \
199
210
--build-arg spark_version="${SPARK_VERSION}" \
200
211
--build-arg jupyterlab_version="${JUPYTERLAB_VERSION}" \
212
+ --build-arg scala_kernel_version="${SCALA_KERNEL_VERSION}" \
201
213
-f docker/jupyterlab/Dockerfile \
202
214
-t jupyterlab:${JUPYTERLAB_VERSION}-spark-${SPARK_VERSION} .
203
215
echo ${DOCKERHUB_PWD} | docker login -u ${DOCKERHUB_USR} --password-stdin
@@ -206,16 +218,30 @@ jobs:
206
218
docker tag jupyterlab:${JUPYTERLAB_VERSION}-spark-${SPARK_VERSION} ${DOCKERHUB_USR}/jupyterlab:latest
207
219
docker push ${DOCKERHUB_USR}/jupyterlab:latest
208
220
221
+ - name : Build Base Image - Scala v2.11.12
222
+ env :
223
+ SCALA_VERSION : " 2.11.12"
224
+ run : |
225
+ cd ${GITHUB_WORKSPACE}/build
226
+ docker build \
227
+ --build-arg build_date="$(date -u +'%Y-%m-%d')" \
228
+ --build-arg scala_version="${SCALA_VERSION}" \
229
+ -f docker/base/Dockerfile \
230
+ -t base:latest .
231
+
209
232
- name : Build & Push JupyterLab Image - Spark v2.4.4
210
233
env :
211
234
SPARK_VERSION : " 2.4.4"
235
+ SCALA_VERSION : " 2.11.12"
236
+ SCALA_KERNEL_VERSION : " 0.6.0"
212
237
run : |
213
238
cd ${GITHUB_WORKSPACE}/build
214
239
docker build \
215
240
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
216
241
--build-arg scala_version="${SCALA_VERSION}" \
217
242
--build-arg spark_version="${SPARK_VERSION}" \
218
243
--build-arg jupyterlab_version="${JUPYTERLAB_VERSION}" \
244
+ --build-arg scala_kernel_version="${SCALA_KERNEL_VERSION}" \
219
245
-f docker/jupyterlab/Dockerfile \
220
246
-t jupyterlab:${JUPYTERLAB_VERSION}-spark-${SPARK_VERSION} .
221
247
echo ${DOCKERHUB_PWD} | docker login -u ${DOCKERHUB_USR} --password-stdin
@@ -225,13 +251,16 @@ jobs:
225
251
- name : Build & Push JupyterLab Image - Spark v2.4.0
226
252
env :
227
253
SPARK_VERSION : " 2.4.0"
254
+ SCALA_VERSION : " 2.11.12"
255
+ SCALA_KERNEL_VERSION : " 0.6.0"
228
256
run : |
229
257
cd ${GITHUB_WORKSPACE}/build
230
258
docker build \
231
259
--build-arg build_date="$(date -u +'%Y-%m-%d')" \
232
260
--build-arg scala_version="${SCALA_VERSION}" \
233
261
--build-arg spark_version="${SPARK_VERSION}" \
234
262
--build-arg jupyterlab_version="${JUPYTERLAB_VERSION}" \
263
+ --build-arg scala_kernel_version="${SCALA_KERNEL_VERSION}" \
235
264
-f docker/jupyterlab/Dockerfile \
236
265
-t jupyterlab:${JUPYTERLAB_VERSION}-spark-${SPARK_VERSION} .
237
266
echo ${DOCKERHUB_PWD} | docker login -u ${DOCKERHUB_USR} --password-stdin
0 commit comments