Skip to content

Commit 229c7fe

Browse files
authored
Merge pull request #1115 from Bidek56/miniconda-py38
Miniconda py38 and Spark 3.0
2 parents 5446280 + 391fdcc commit 229c7fe

File tree

6 files changed

+15
-113
lines changed

6 files changed

+15
-113
lines changed

Makefile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -88,7 +88,7 @@ lint-build-test-all: $(foreach I,$(ALL_IMAGES),lint/$(I) arch_patch/$(I) build/$
8888

8989
lint-install: ## install hadolint
9090
@echo "Installing hadolint at $(HADOLINT) ..."
91-
@curl -sL -o $(HADOLINT) "https://github.com/hadolint/hadolint/releases/download/v1.17.6/hadolint-$(shell uname -s)-$(shell uname -m)"
91+
@curl -sL -o $(HADOLINT) "https://github.com/hadolint/hadolint/releases/download/v1.18.0/hadolint-$(shell uname -s)-$(shell uname -m)"
9292
@chmod 700 $(HADOLINT)
9393
@echo "Installation done!"
9494
@$(HADOLINT) --version

all-spark-notebook/Dockerfile

Lines changed: 0 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -33,16 +33,6 @@ RUN conda install --quiet --yes \
3333
fix-permissions "${CONDA_DIR}" && \
3434
fix-permissions "/home/${NB_USER}"
3535

36-
# Apache Toree kernel
37-
# hadolint ignore=DL3013
38-
RUN pip install --no-cache-dir \
39-
https://dist.apache.org/repos/dist/release/incubator/toree/0.3.0-incubating/toree-pip/toree-0.3.0.tar.gz \
40-
&& \
41-
jupyter toree install --sys-prefix && \
42-
rm -rf "/home/${NB_USER}/.local" && \
43-
fix-permissions "${CONDA_DIR}" && \
44-
fix-permissions "/home/${NB_USER}"
45-
4636
# Spylon-kernel
4737
RUN conda install --quiet --yes 'spylon-kernel=0.4*' && \
4838
conda clean --all -f -y && \

all-spark-notebook/test/data/local_toree.ipynb

Lines changed: 0 additions & 89 deletions
This file was deleted.

all-spark-notebook/test/test_spark_notebooks.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
@pytest.mark.parametrize(
1313
"test_file",
1414
# TODO: add local_sparklyr
15-
["local_pyspark", "local_spylon", "local_toree", "local_sparkR"],
15+
["local_pyspark", "local_spylon", "local_sparkR"],
1616
)
1717
def test_nbconvert(container, test_file):
1818
"""Check if Spark notebooks can be executed"""

base-notebook/Dockerfile

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -76,14 +76,14 @@ RUN mkdir /home/$NB_USER/work && \
7676

7777
# Install conda as jovyan and check the md5 sum provided on the download site
7878
ENV MINICONDA_VERSION=4.8.2 \
79-
MINICONDA_MD5=87e77f097f6ebb5127c77662dfc3165e \
79+
MINICONDA_MD5=cbda751e713b5a95f187ae70b509403f \
8080
CONDA_VERSION=4.8.2
8181

8282
WORKDIR /tmp
83-
RUN wget --quiet https://repo.continuum.io/miniconda/Miniconda3-py37_${MINICONDA_VERSION}-Linux-x86_64.sh && \
84-
echo "${MINICONDA_MD5} *Miniconda3-py37_${MINICONDA_VERSION}-Linux-x86_64.sh" | md5sum -c - && \
85-
/bin/bash Miniconda3-py37_${MINICONDA_VERSION}-Linux-x86_64.sh -f -b -p $CONDA_DIR && \
86-
rm Miniconda3-py37_${MINICONDA_VERSION}-Linux-x86_64.sh && \
83+
RUN wget --quiet https://repo.continuum.io/miniconda/Miniconda3-py38_${MINICONDA_VERSION}-Linux-x86_64.sh && \
84+
echo "${MINICONDA_MD5} *Miniconda3-py38_${MINICONDA_VERSION}-Linux-x86_64.sh" | md5sum -c - && \
85+
/bin/bash Miniconda3-py38_${MINICONDA_VERSION}-Linux-x86_64.sh -f -b -p $CONDA_DIR && \
86+
rm Miniconda3-py38_${MINICONDA_VERSION}-Linux-x86_64.sh && \
8787
echo "conda ${CONDA_VERSION}" >> $CONDA_DIR/conda-meta/pinned && \
8888
conda config --system --prepend channels conda-forge && \
8989
conda config --system --set auto_update_conda false && \
@@ -115,7 +115,7 @@ RUN conda install --quiet --yes 'tini=0.18.0' && \
115115
RUN conda install --quiet --yes \
116116
'notebook=6.0.3' \
117117
'jupyterhub=1.1.0' \
118-
'jupyterlab=2.1.3' && \
118+
'jupyterlab=2.1.4' && \
119119
conda clean --all -f -y && \
120120
npm cache clean --force && \
121121
jupyter notebook --generate-config && \

pyspark-notebook/Dockerfile

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -11,19 +11,20 @@ SHELL ["/bin/bash", "-o", "pipefail", "-c"]
1111
USER root
1212

1313
# Spark dependencies
14-
ENV APACHE_SPARK_VERSION=2.4.5 \
15-
HADOOP_VERSION=2.7
14+
ENV APACHE_SPARK_VERSION=3.0.0 \
15+
HADOOP_VERSION=3.2
1616

1717
RUN apt-get -y update && \
18-
apt-get install --no-install-recommends -y openjdk-8-jre-headless ca-certificates-java && \
18+
apt-get install --no-install-recommends -y openjdk-11-jre-headless ca-certificates-java && \
1919
rm -rf /var/lib/apt/lists/*
2020

2121
# Using the preferred mirror to download Spark
2222
WORKDIR /tmp
23+
2324
# hadolint ignore=SC2046
2425
RUN wget -q $(wget -qO- https://www.apache.org/dyn/closer.lua/spark/spark-${APACHE_SPARK_VERSION}/spark-${APACHE_SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz\?as_json | \
2526
python -c "import sys, json; content=json.load(sys.stdin); print(content['preferred']+content['path_info'])") && \
26-
echo "2426a20c548bdfc07df288cd1d18d1da6b3189d0b78dee76fa034c52a4e02895f0ad460720c526f163ba63a17efae4764c46a1cd8f9b04c60f9937a554db85d2 *spark-${APACHE_SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz" | sha512sum -c - && \
27+
echo "BFE45406C67CC4AE00411AD18CC438F51E7D4B6F14EB61E7BF6B5450897C2E8D3AB020152657C0239F253735C263512FFABF538AC5B9FFFA38B8295736A9C387 *spark-${APACHE_SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz" | sha512sum -c - && \
2728
tar xzf "spark-${APACHE_SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz" -C /usr/local --owner root --group root --no-same-owner && \
2829
rm "spark-${APACHE_SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz"
2930

@@ -32,7 +33,7 @@ RUN ln -s "spark-${APACHE_SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}" spark
3233

3334
# Configure Spark
3435
ENV SPARK_HOME=/usr/local/spark
35-
ENV PYTHONPATH=$SPARK_HOME/python:$SPARK_HOME/python/lib/py4j-0.10.7-src.zip \
36+
ENV PYTHONPATH=$SPARK_HOME/python:$SPARK_HOME/python/lib/py4j-0.10.9-src.zip \
3637
SPARK_OPTS="--driver-java-options=-Xms1024M --driver-java-options=-Xmx4096M --driver-java-options=-Dlog4j.logLevel=info" \
3738
PATH=$PATH:$SPARK_HOME/bin
3839

@@ -44,4 +45,4 @@ RUN conda install --quiet -y 'pyarrow' && \
4445
fix-permissions "${CONDA_DIR}" && \
4546
fix-permissions "/home/${NB_USER}"
4647

47-
WORKDIR $HOME
48+
WORKDIR $HOME

0 commit comments

Comments
 (0)