diff --git a/examples/spark/images/Makefile b/examples/spark/images/Makefile deleted file mode 100644 index 9d465e228b1..00000000000 --- a/examples/spark/images/Makefile +++ /dev/null @@ -1,61 +0,0 @@ -all: spark zeppelin -push: push-spark push-zeppelin -.PHONY: push push-spark push-zeppelin spark zeppelin zeppelin-build - -# To bump the Spark version, bump the version in base/Dockerfile, bump -# the version in zeppelin/Dockerfile, bump this tag and reset to -# v1. You should also double check the native Hadoop libs at that -# point (we grab the 2.6.1 libs, which are appropriate for -# 1.5.1-with-2.6). Note that you'll need to re-test Zeppelin (and it -# may not have caught up to newest Spark). -TAG = 1.5.1_v2 - -# To bump the Zeppelin version, bump the version in -# zeppelin/Dockerfile and bump this tag and reset to v1. -ZEPPELIN_TAG = v0.5.5_v3 - -spark: - docker build -t gcr.io/google_containers/spark-base base - docker tag gcr.io/google_containers/spark-base gcr.io/google_containers/spark-base:$(TAG) - docker build -t gcr.io/google_containers/spark-worker worker - docker tag gcr.io/google_containers/spark-worker gcr.io/google_containers/spark-worker:$(TAG) - docker build -t gcr.io/google_containers/spark-master master - docker tag gcr.io/google_containers/spark-master gcr.io/google_containers/spark-master:$(TAG) - docker build -t gcr.io/google_containers/spark-driver driver - docker tag gcr.io/google_containers/spark-driver gcr.io/google_containers/spark-driver:$(TAG) - -# This target is useful when needing to use an unreleased version of Zeppelin -zeppelin-build: - docker build -t gcr.io/google_containers/zeppelin-build zeppelin-build - docker tag -f gcr.io/google_containers/zeppelin-build gcr.io/google_containers/zeppelin-build:$(ZEPPELIN_TAG) - -zeppelin: - docker build -t gcr.io/google_containers/zeppelin zeppelin - docker tag -f gcr.io/google_containers/zeppelin gcr.io/google_containers/zeppelin:$(ZEPPELIN_TAG) - -push-spark: spark - gcloud docker push gcr.io/google_containers/spark-base - gcloud docker push gcr.io/google_containers/spark-base:$(TAG) - gcloud docker push gcr.io/google_containers/spark-worker - gcloud docker push gcr.io/google_containers/spark-worker:$(TAG) - gcloud docker push gcr.io/google_containers/spark-master - gcloud docker push gcr.io/google_containers/spark-master:$(TAG) - gcloud docker push gcr.io/google_containers/spark-driver - gcloud docker push gcr.io/google_containers/spark-driver:$(TAG) - -push-zeppelin: zeppelin - gcloud docker push gcr.io/google_containers/zeppelin - gcloud docker push gcr.io/google_containers/zeppelin:$(ZEPPELIN_TAG) - -clean: - docker rmi gcr.io/google_containers/spark-worker:$(TAG) || : - docker rmi gcr.io/google_containers/spark-worker || : - docker rmi gcr.io/google_containers/spark-master:$(TAG) || : - docker rmi gcr.io/google_containers/spark-master || : - docker rmi gcr.io/google_containers/spark-driver:$(TAG) || : - docker rmi gcr.io/google_containers/spark-driver || : - docker rmi gcr.io/google_containers/spark-base:$(TAG) || : - docker rmi gcr.io/google_containers/spark-base || : - - docker rmi gcr.io/google_containers/zeppelin:$(ZEPPELIN_TAG) || : - docker rmi gcr.io/google_containers/zeppelin || : diff --git a/examples/spark/images/base/Dockerfile b/examples/spark/images/base/Dockerfile deleted file mode 100644 index 647defeaaaf..00000000000 --- a/examples/spark/images/base/Dockerfile +++ /dev/null @@ -1,39 +0,0 @@ -FROM java:openjdk-8-jdk - -ENV hadoop_ver 2.6.1 -ENV spark_ver 1.5.1 - -# Get Hadoop from US Apache mirror and extract just the native -# libs. (Until we care about running HDFS with these containers, this -# is all we need.) -RUN mkdir -p /opt && \ - cd /opt && \ - curl http://www.us.apache.org/dist/hadoop/common/hadoop-${hadoop_ver}/hadoop-${hadoop_ver}.tar.gz | \ - tar -zx hadoop-${hadoop_ver}/lib/native && \ - ln -s hadoop-${hadoop_ver} hadoop && \ - echo Hadoop ${hadoop_ver} native libraries installed in /opt/hadoop/lib/native - -# Get Spark from US Apache mirror. -RUN mkdir -p /opt && \ - cd /opt && \ - curl http://www.us.apache.org/dist/spark/spark-${spark_ver}/spark-${spark_ver}-bin-hadoop2.6.tgz | \ - tar -zx && \ - ln -s spark-${spark_ver}-bin-hadoop2.6 spark && \ - echo Spark ${spark_ver} installed in /opt - -# Add the GCS connector. -RUN cd /opt/spark/lib && \ - curl -O https://storage.googleapis.com/hadoop-lib/gcs/gcs-connector-latest-hadoop2.jar - -# if numpy is installed on a driver it needs to be installed on all -# workers, so install it everywhere -RUN apt-get update && \ - apt-get install -y python-numpy && \ - apt-get clean && \ - rm -rf /var/lib/apt/lists/* - -ADD log4j.properties /opt/spark/conf/log4j.properties -ADD start-common.sh / -ADD core-site.xml /opt/spark/conf/core-site.xml -ADD spark-defaults.conf /opt/spark/conf/spark-defaults.conf -ENV PATH $PATH:/opt/spark/bin diff --git a/examples/spark/images/base/core-site.xml b/examples/spark/images/base/core-site.xml deleted file mode 100644 index 2fecabedc8f..00000000000 --- a/examples/spark/images/base/core-site.xml +++ /dev/null @@ -1,19 +0,0 @@ - - - - - - fs.gs.impl - com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystem - The FileSystem for gs: (GCS) uris. - - - fs.AbstractFileSystem.gs.impl - com.google.cloud.hadoop.fs.gcs.GoogleHadoopFS - The AbstractFileSystem for gs: (GCS) uris. Only necessary for use with Hadoop 2. - - - fs.gs.project.id - NOT_RUNNING_INSIDE_GCE - - diff --git a/examples/spark/images/base/log4j.properties b/examples/spark/images/base/log4j.properties deleted file mode 100644 index b146f8a7841..00000000000 --- a/examples/spark/images/base/log4j.properties +++ /dev/null @@ -1,12 +0,0 @@ -# Set everything to be logged to the console -log4j.rootCategory=WARN, console -log4j.appender.console=org.apache.log4j.ConsoleAppender -log4j.appender.console.target=System.err -log4j.appender.console.layout=org.apache.log4j.PatternLayout -log4j.appender.console.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{1}: %m%n - -# Settings to quiet third party logs that are too verbose -log4j.logger.org.spark-project.jetty=WARN -log4j.logger.org.spark-project.jetty.util.component.AbstractLifeCycle=ERROR -log4j.logger.org.apache.spark.repl.SparkIMain$exprTyper=INFO -log4j.logger.org.apache.spark.repl.SparkILoop$SparkILoopInterpreter=INFO diff --git a/examples/spark/images/base/spark-defaults.conf b/examples/spark/images/base/spark-defaults.conf deleted file mode 100644 index f5d3d98504c..00000000000 --- a/examples/spark/images/base/spark-defaults.conf +++ /dev/null @@ -1,5 +0,0 @@ -spark.master spark://spark-master:7077 -spark.executor.extraClassPath /opt/spark/lib/gcs-connector-latest-hadoop2.jar -spark.driver.extraClassPath /opt/spark/lib/gcs-connector-latest-hadoop2.jar -spark.driver.extraLibraryPath /opt/hadoop/lib/native -spark.app.id KubernetesSpark diff --git a/examples/spark/images/base/start-common.sh b/examples/spark/images/base/start-common.sh deleted file mode 100755 index ac8d5058386..00000000000 --- a/examples/spark/images/base/start-common.sh +++ /dev/null @@ -1,30 +0,0 @@ -#!/bin/bash - -# Copyright 2015 The Kubernetes Authors All rights reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -PROJECT_ID=$(curl -s -H "Metadata-Flavor: Google" http://metadata.google.internal/computeMetadata/v1/project/project-id) - -if [[ -n "${PROJECT_ID}" ]]; then - sed -i "s/NOT_RUNNING_INSIDE_GCE/${PROJECT_ID}/" /opt/spark/conf/core-site.xml -fi - -# We don't want any of the incoming service variables, we'd rather use -# DNS. But this one interferes directly with Spark. -unset SPARK_MASTER_PORT - -# spark.{executor,driver}.extraLibraryPath don't actually seem to -# work, this seems to be the only reliable way to get the native libs -# picked up. -export LD_LIBRARY_PATH=${LD_LIBRARY_PATH}:/opt/hadoop/lib/native diff --git a/examples/spark/images/driver/Dockerfile b/examples/spark/images/driver/Dockerfile deleted file mode 100644 index 7dc33857321..00000000000 --- a/examples/spark/images/driver/Dockerfile +++ /dev/null @@ -1,3 +0,0 @@ -FROM gcr.io/google_containers/spark-base:latest -ADD start.sh /start.sh -CMD ["/start.sh"] diff --git a/examples/spark/images/driver/README.md b/examples/spark/images/driver/README.md deleted file mode 100644 index a418e5dfa06..00000000000 --- a/examples/spark/images/driver/README.md +++ /dev/null @@ -1,38 +0,0 @@ - - - - -WARNING -WARNING -WARNING -WARNING -WARNING - -

PLEASE NOTE: This document applies to the HEAD of the source tree

- -If you are using a released version of Kubernetes, you should -refer to the docs that go with that version. - - - -The latest release of this document can be found -[here](http://releases.k8s.io/release-1.1/examples/spark/images/driver/README.md). - -Documentation for other releases can be found at -[releases.k8s.io](http://releases.k8s.io). - --- - - - - - - - -[![Analytics](https://kubernetes-site.appspot.com/UA-36037335-10/GitHub/examples/spark/images/driver/README.md?pixel)]() - diff --git a/examples/spark/images/driver/start.sh b/examples/spark/images/driver/start.sh deleted file mode 100755 index 89d4fcec00f..00000000000 --- a/examples/spark/images/driver/start.sh +++ /dev/null @@ -1,22 +0,0 @@ -#!/bin/bash - -# Copyright 2015 The Kubernetes Authors All rights reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -. /start-common.sh - -echo "Use kubectl exec spark-driver -it bash to invoke commands" -while true; do - sleep 100 -done diff --git a/examples/spark/images/master/Dockerfile b/examples/spark/images/master/Dockerfile deleted file mode 100644 index d63cc29de7f..00000000000 --- a/examples/spark/images/master/Dockerfile +++ /dev/null @@ -1,7 +0,0 @@ -FROM gcr.io/google_containers/spark-base:latest - -ADD start.sh / -ADD log4j.properties /opt/spark/conf/log4j.properties -EXPOSE 7077 8080 - -ENTRYPOINT ["/start.sh"] diff --git a/examples/spark/images/master/log4j.properties b/examples/spark/images/master/log4j.properties deleted file mode 100644 index 3a2a8821981..00000000000 --- a/examples/spark/images/master/log4j.properties +++ /dev/null @@ -1,12 +0,0 @@ -# Set everything to be logged to the console -log4j.rootCategory=INFO, console -log4j.appender.console=org.apache.log4j.ConsoleAppender -log4j.appender.console.target=System.err -log4j.appender.console.layout=org.apache.log4j.PatternLayout -log4j.appender.console.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{1}: %m%n - -# Settings to quiet third party logs that are too verbose -log4j.logger.org.spark-project.jetty=WARN -log4j.logger.org.spark-project.jetty.util.component.AbstractLifeCycle=ERROR -log4j.logger.org.apache.spark.repl.SparkIMain$exprTyper=INFO -log4j.logger.org.apache.spark.repl.SparkILoop$SparkILoopInterpreter=INFO diff --git a/examples/spark/images/master/start.sh b/examples/spark/images/master/start.sh deleted file mode 100755 index 7592c8a5278..00000000000 --- a/examples/spark/images/master/start.sh +++ /dev/null @@ -1,22 +0,0 @@ -#!/bin/bash - -# Copyright 2015 The Kubernetes Authors All rights reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -. /start-common.sh - -echo "$(hostname -i) spark-master" >> /etc/hosts - -# Run spark-class directly so that when it exits (or crashes), the pod restarts. -/opt/spark/bin/spark-class org.apache.spark.deploy.master.Master --ip spark-master --port 7077 --webui-port 8080 diff --git a/examples/spark/images/worker/Dockerfile b/examples/spark/images/worker/Dockerfile deleted file mode 100644 index bb113e2dbd4..00000000000 --- a/examples/spark/images/worker/Dockerfile +++ /dev/null @@ -1,7 +0,0 @@ -FROM gcr.io/google_containers/spark-base:latest - -ADD start.sh / -ADD log4j.properties /opt/spark/conf/log4j.properties -EXPOSE 8081 - -ENTRYPOINT ["/start.sh"] diff --git a/examples/spark/images/worker/log4j.properties b/examples/spark/images/worker/log4j.properties deleted file mode 100644 index 3a2a8821981..00000000000 --- a/examples/spark/images/worker/log4j.properties +++ /dev/null @@ -1,12 +0,0 @@ -# Set everything to be logged to the console -log4j.rootCategory=INFO, console -log4j.appender.console=org.apache.log4j.ConsoleAppender -log4j.appender.console.target=System.err -log4j.appender.console.layout=org.apache.log4j.PatternLayout -log4j.appender.console.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{1}: %m%n - -# Settings to quiet third party logs that are too verbose -log4j.logger.org.spark-project.jetty=WARN -log4j.logger.org.spark-project.jetty.util.component.AbstractLifeCycle=ERROR -log4j.logger.org.apache.spark.repl.SparkIMain$exprTyper=INFO -log4j.logger.org.apache.spark.repl.SparkILoop$SparkILoopInterpreter=INFO diff --git a/examples/spark/images/worker/start.sh b/examples/spark/images/worker/start.sh deleted file mode 100755 index 058df5b18a9..00000000000 --- a/examples/spark/images/worker/start.sh +++ /dev/null @@ -1,20 +0,0 @@ -#!/bin/bash - -# Copyright 2015 The Kubernetes Authors All rights reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -. /start-common.sh - -# Run spark-class directly so that when it exits (or crashes), the pod restarts. -/opt/spark/bin/spark-class org.apache.spark.deploy.worker.Worker spark://spark-master:7077 --webui-port 8081 diff --git a/examples/spark/images/zeppelin-build/Dockerfile b/examples/spark/images/zeppelin-build/Dockerfile deleted file mode 100644 index 01f9c909e8a..00000000000 --- a/examples/spark/images/zeppelin-build/Dockerfile +++ /dev/null @@ -1,54 +0,0 @@ -# Copyright 2015 The Kubernetes Authors All rights reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -# This is the Zeppelin *build* image. It spits out a /zeppelin.tgz -# alone, which is then copied out by the Makefile and used in the -# actual Zeppelin image. -# -# Based heavily on -# https://github.com/dylanmei/docker-zeppelin/blob/master/Dockerfile -# (which is similar to many others out there), but rebased onto maven -# image. - -FROM maven:3.3.3-jdk-8 - -ENV ZEPPELIN_TAG v0.5.5 -ENV SPARK_MINOR 1.5 -ENV SPARK_PATCH 1 -ENV SPARK_VER ${SPARK_MINOR}.${SPARK_PATCH} -ENV HADOOP_MINOR 2.6 -ENV HADOOP_PATCH 1 -ENV HADOOP_VER ${HADOOP_MINOR}.${HADOOP_PATCH} - -# libfontconfig is a workaround for -# https://github.com/karma-runner/karma/issues/1270, which caused a -# build break similar to -# https://www.mail-archive.com/users@zeppelin.incubator.apache.org/msg01586.html - -RUN apt-get update \ - && apt-get install -y net-tools build-essential git wget unzip python python-setuptools python-dev python-numpy libfontconfig - -RUN git clone https://github.com/apache/incubator-zeppelin.git --branch ${ZEPPELIN_TAG} /opt/zeppelin -RUN cd /opt/zeppelin && \ - mvn clean package \ - -Pbuild-distr \ - -Pspark-${SPARK_MINOR} -Dspark.version=${SPARK_VER} \ - -Phadoop-${HADOOP_MINOR} -Dhadoop.version=${HADOOP_VER} \ - -Ppyspark \ - -DskipTests && \ - echo "Successfully built Zeppelin" - -RUN cd /opt/zeppelin/zeppelin-distribution/target/zeppelin-* && \ - mv zeppelin-* zeppelin && \ - tar cvzf /zeppelin.tgz zeppelin diff --git a/examples/spark/images/zeppelin/.gitignore b/examples/spark/images/zeppelin/.gitignore deleted file mode 100644 index 9fae9b4cafe..00000000000 --- a/examples/spark/images/zeppelin/.gitignore +++ /dev/null @@ -1 +0,0 @@ -zeppelin.tgz diff --git a/examples/spark/images/zeppelin/Dockerfile b/examples/spark/images/zeppelin/Dockerfile deleted file mode 100644 index 0208766e864..00000000000 --- a/examples/spark/images/zeppelin/Dockerfile +++ /dev/null @@ -1,33 +0,0 @@ -# Copyright 2015 The Kubernetes Authors All rights reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -# This image relies on the zeppelin-build image to get the zeppelin -# binaries built, and the Makefile to put it in this directory. - -FROM gcr.io/google_containers/spark-base:latest - -ENV ZEPPELIN_VER 0.5.5-incubating - -RUN mkdir -p /opt && \ - cd /opt && \ - curl http://www.us.apache.org/dist/incubator/zeppelin/${ZEPPELIN_VER}/zeppelin-${ZEPPELIN_VER}-bin-all.tgz | \ - tar -zx && \ - ln -s zeppelin-${ZEPPELIN_VER}-bin-all zeppelin && \ - echo Zeppelin ${ZEPPELIN_VER} installed in /opt - -ADD zeppelin-log4j.properties /opt/zeppelin/conf/log4j.properties -ADD zeppelin-env.sh /opt/zeppelin/conf/zeppelin-env.sh -ADD docker-zeppelin.sh /opt/zeppelin/bin/docker-zeppelin.sh -EXPOSE 8080 -ENTRYPOINT ["/opt/zeppelin/bin/docker-zeppelin.sh"] diff --git a/examples/spark/images/zeppelin/docker-zeppelin.sh b/examples/spark/images/zeppelin/docker-zeppelin.sh deleted file mode 100755 index 99cbde1660c..00000000000 --- a/examples/spark/images/zeppelin/docker-zeppelin.sh +++ /dev/null @@ -1,21 +0,0 @@ -#!/bin/bash - -# Copyright 2015 The Kubernetes Authors All rights reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -export ZEPPELIN_HOME=/opt/zeppelin -export ZEPPELIN_CONF_DIR="${ZEPPELIN_HOME}/conf" - -echo "=== Launching Zeppelin under Docker ===" -/opt/zeppelin/bin/zeppelin.sh "${ZEPPELIN_CONF_DIR}" diff --git a/examples/spark/images/zeppelin/zeppelin-env.sh b/examples/spark/images/zeppelin/zeppelin-env.sh deleted file mode 100644 index 3fa8a024867..00000000000 --- a/examples/spark/images/zeppelin/zeppelin-env.sh +++ /dev/null @@ -1,26 +0,0 @@ -#!/bin/bash - -# Copyright 2015 The Kubernetes Authors All rights reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -export MASTER="spark://spark-master:7077" -export SPARK_HOME=/opt/spark -export ZEPPELIN_JAVA_OPTS="-Dspark.jars=/opt/spark/lib/gcs-connector-latest-hadoop2.jar" -# TODO(zmerlynn): Setting global CLASSPATH *should* be unnecessary, -# but ZEPPELIN_JAVA_OPTS isn't enough here. :( -export CLASSPATH="/opt/spark/lib/gcs-connector-latest-hadoop2.jar" -export ZEPPELIN_NOTEBOOK_DIR="${ZEPPELIN_HOME}/notebook" -export ZEPPELIN_MEM=-Xmx1024m -export ZEPPELIN_PORT=8080 -export PYTHONPATH="${SPARK_HOME}/python:${SPARK_HOME}/python/lib/py4j-0.8.2.1-src.zip" diff --git a/examples/spark/images/zeppelin/zeppelin-log4j.properties b/examples/spark/images/zeppelin/zeppelin-log4j.properties deleted file mode 100644 index 5fb18104eed..00000000000 --- a/examples/spark/images/zeppelin/zeppelin-log4j.properties +++ /dev/null @@ -1,6 +0,0 @@ -# Set everything to be logged to the console. -log4j.rootCategory=INFO, console -log4j.appender.console=org.apache.log4j.ConsoleAppender -log4j.appender.console.target=System.err -log4j.appender.console.layout=org.apache.log4j.PatternLayout -log4j.appender.console.layout.ConversionPattern=%5p [%d] ({%t} %F[%M]:%L) - %m%n