增加spark-on-kubernetes需要用到的yaml文件和镜像编译脚本
parent
0bbc871650
commit
6577f18131
|
@ -0,0 +1,72 @@
|
|||
#!/usr/bin/env bash
|
||||
|
||||
# Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
# contributor license agreements. See the NOTICE file distributed with
|
||||
# this work for additional information regarding copyright ownership.
|
||||
# The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
# (the "License"); you may not use this file except in compliance with
|
||||
# the License. You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
#
|
||||
|
||||
# This script builds and pushes docker images when run from a release of Spark
|
||||
# with Kubernetes support.
|
||||
|
||||
declare -A path=( [spark-driver]=dockerfiles/driver/Dockerfile \
|
||||
[spark-executor]=dockerfiles/executor/Dockerfile \
|
||||
[spark-driver-py]=dockerfiles/driver-py/Dockerfile \
|
||||
[spark-executor-py]=dockerfiles/executor-py/Dockerfile \
|
||||
[spark-init]=dockerfiles/init-container/Dockerfile \
|
||||
[spark-shuffle]=dockerfiles/shuffle-service/Dockerfile \
|
||||
[spark-resource-staging-server]=dockerfiles/resource-staging-server/Dockerfile )
|
||||
|
||||
function build {
|
||||
docker build -t spark-base -f dockerfiles/spark-base/Dockerfile .
|
||||
for image in "${!path[@]}"; do
|
||||
docker build -t ${REPO}/$image:${TAG} -f ${path[$image]} .
|
||||
done
|
||||
}
|
||||
|
||||
|
||||
function push {
|
||||
for image in "${!path[@]}"; do
|
||||
docker push ${REPO}/$image:${TAG}
|
||||
done
|
||||
}
|
||||
|
||||
function usage {
|
||||
echo "Usage: ./sbin/build-push-docker-images.sh -r <repo> -t <tag> build"
|
||||
echo " ./sbin/build-push-docker-images.sh -r <repo> -t <tag> push"
|
||||
echo "for example: ./sbin/build-push-docker-images.sh -r docker.io/kubespark -t v2.2.0 push"
|
||||
}
|
||||
|
||||
if [[ "$@" = *--help ]] || [[ "$@" = *-h ]]; then
|
||||
usage
|
||||
exit 0
|
||||
fi
|
||||
|
||||
while getopts r:t: option
|
||||
do
|
||||
case "${option}"
|
||||
in
|
||||
r) REPO=${OPTARG};;
|
||||
t) TAG=${OPTARG};;
|
||||
esac
|
||||
done
|
||||
|
||||
if [ -z "$REPO" ] || [ -z "$TAG" ]; then
|
||||
usage
|
||||
else
|
||||
case "${@: -1}" in
|
||||
build) build;;
|
||||
push) push;;
|
||||
*) usage;;
|
||||
esac
|
||||
fi
|
|
@ -0,0 +1,82 @@
|
|||
#
|
||||
# Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
# contributor license agreements. See the NOTICE file distributed with
|
||||
# this work for additional information regarding copyright ownership.
|
||||
# The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
# (the "License"); you may not use this file except in compliance with
|
||||
# the License. You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
#
|
||||
---
|
||||
apiVersion: extensions/v1beta1
|
||||
kind: Deployment
|
||||
metadata:
|
||||
name: spark-resource-staging-server
|
||||
namespace: spark-cluster
|
||||
spec:
|
||||
replicas: 1
|
||||
template:
|
||||
metadata:
|
||||
labels:
|
||||
resource-staging-server-instance: default
|
||||
spec:
|
||||
volumes:
|
||||
- name: resource-staging-server-properties
|
||||
configMap:
|
||||
name: spark-resource-staging-server-config
|
||||
containers:
|
||||
- name: spark-resource-staging-server
|
||||
image: sz-pg-oam-docker-hub-001.tendcloud.com/library/spark-resource-staging-server:v2.1.0-kubernetes-0.3.1-1
|
||||
resources:
|
||||
requests:
|
||||
cpu: 100m
|
||||
memory: 256Mi
|
||||
limits:
|
||||
cpu: 1000m
|
||||
memory: 2560Mi
|
||||
volumeMounts:
|
||||
- name: resource-staging-server-properties
|
||||
mountPath: '/etc/spark-resource-staging-server'
|
||||
args:
|
||||
- '/etc/spark-resource-staging-server/resource-staging-server.properties'
|
||||
---
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: spark-resource-staging-server-config
|
||||
namespace: spark-cluster
|
||||
data:
|
||||
resource-staging-server.properties: |
|
||||
spark.kubernetes.resourceStagingServer.port=10000
|
||||
spark.ssl.kubernetes.resourceStagingServer.enabled=false
|
||||
# Other possible properties are listed below, primarily for setting up TLS. The paths given by KeyStore, password, and PEM files here should correspond to
|
||||
# files that are securely mounted into the resource staging server container, via e.g. secret volumes.
|
||||
# spark.ssl.kubernetes.resourceStagingServer.keyStore=/mnt/secrets/resource-staging-server/keyStore.jks
|
||||
# spark.ssl.kubernetes.resourceStagingServer.keyStorePassword=changeit
|
||||
# spark.ssl.kubernetes.resourceStagingServer.keyPassword=changeit
|
||||
# spark.ssl.kubernetes.resourceStagingServer.keyStorePasswordFile=/mnt/secrets/resource-staging-server/keystore-password.txt
|
||||
# spark.ssl.kubernetes.resourceStagingServer.keyPasswordFile=/mnt/secrets/resource-staging-server/keystore-key-password.txt
|
||||
# spark.ssl.kubernetes.resourceStagingServer.keyPem=/mnt/secrets/resource-staging-server/key.pem
|
||||
# spark.ssl.kubernetes.resourceStagingServer.serverCertPem=/mnt/secrets/resource-staging-server/cert.pem
|
||||
---
|
||||
apiVersion: v1
|
||||
kind: Service
|
||||
metadata:
|
||||
name: spark-resource-staging-service
|
||||
namespace: spark-cluster
|
||||
spec:
|
||||
type: NodePort
|
||||
selector:
|
||||
resource-staging-server-instance: default
|
||||
ports:
|
||||
- protocol: TCP
|
||||
port: 10000
|
||||
targetPort: 10000
|
||||
nodePort: 31000
|
Loading…
Reference in New Issue