Skip to content

Commit e92251a

Browse files
author
Yeongil Jung
committed
Update py4j from 0.10.9 to 0.10.9.7
1 parent 72bdeb2 commit e92251a

File tree

13 files changed

+11
-11
lines changed

13 files changed

+11
-11
lines changed

bin/pyspark

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ export PYSPARK_DRIVER_PYTHON_OPTS
5050

5151
# Add the PySpark classes to the Python path:
5252
export PYTHONPATH="${SPARK_HOME}/python/:$PYTHONPATH"
53-
export PYTHONPATH="${SPARK_HOME}/python/lib/py4j-0.10.9-src.zip:$PYTHONPATH"
53+
export PYTHONPATH="${SPARK_HOME}/python/lib/py4j-0.10.9.7-src.zip:$PYTHONPATH"
5454

5555
# Load the PySpark shell.py script when ./pyspark is used interactively:
5656
export OLD_PYTHONSTARTUP="$PYTHONSTARTUP"

bin/pyspark2.cmd

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ if "x%PYSPARK_DRIVER_PYTHON%"=="x" (
3030
)
3131

3232
set PYTHONPATH=%SPARK_HOME%\python;%PYTHONPATH%
33-
set PYTHONPATH=%SPARK_HOME%\python\lib\py4j-0.10.9-src.zip;%PYTHONPATH%
33+
set PYTHONPATH=%SPARK_HOME%\python\lib\py4j-0.10.9.7-src.zip;%PYTHONPATH%
3434

3535
set OLD_PYTHONSTARTUP=%PYTHONSTARTUP%
3636
set PYTHONSTARTUP=%SPARK_HOME%\python\pyspark\shell.py

core/pom.xml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -414,7 +414,7 @@
414414
<dependency>
415415
<groupId>net.sf.py4j</groupId>
416416
<artifactId>py4j</artifactId>
417-
<version>0.10.9</version>
417+
<version>0.10.9.7</version>
418418
</dependency>
419419
<dependency>
420420
<groupId>org.apache.spark</groupId>

core/src/main/scala/org/apache/spark/api/python/PythonUtils.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ import org.apache.spark.SparkContext
2727
import org.apache.spark.api.java.{JavaRDD, JavaSparkContext}
2828

2929
private[spark] object PythonUtils {
30-
val PY4J_ZIP_NAME = "py4j-0.10.9-src.zip"
30+
val PY4J_ZIP_NAME = "py4j-0.10.9.7-src.zip"
3131

3232
/** Get the PYTHONPATH for PySpark, either from SPARK_HOME, if it is set, or from our JAR */
3333
def sparkPythonPath: String = {

dev/deps/spark-deps-hadoop-2.7-hive-1.2

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -175,7 +175,7 @@ parquet-hadoop-bundle/1.6.0//parquet-hadoop-bundle-1.6.0.jar
175175
parquet-hadoop/1.10.1//parquet-hadoop-1.10.1.jar
176176
parquet-jackson/1.10.1//parquet-jackson-1.10.1.jar
177177
protobuf-java/2.5.0//protobuf-java-2.5.0.jar
178-
py4j/0.10.9//py4j-0.10.9.jar
178+
py4j/0.10.9.7//py4j-0.10.9.7.jar
179179
pyrolite/4.30//pyrolite-4.30.jar
180180
scala-collection-compat_2.12/2.1.1//scala-collection-compat_2.12-2.1.1.jar
181181
scala-compiler/2.12.10//scala-compiler-2.12.10.jar

dev/deps/spark-deps-hadoop-2.7-hive-2.3

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -189,7 +189,7 @@ parquet-format/2.4.0//parquet-format-2.4.0.jar
189189
parquet-hadoop/1.10.1//parquet-hadoop-1.10.1.jar
190190
parquet-jackson/1.10.1//parquet-jackson-1.10.1.jar
191191
protobuf-java/2.5.0//protobuf-java-2.5.0.jar
192-
py4j/0.10.9//py4j-0.10.9.jar
192+
py4j/0.10.9.7//py4j-0.10.9.7.jar
193193
pyrolite/4.30//pyrolite-4.30.jar
194194
scala-collection-compat_2.12/2.1.1//scala-collection-compat_2.12-2.1.1.jar
195195
scala-compiler/2.12.10//scala-compiler-2.12.10.jar

dev/deps/spark-deps-hadoop-3.2-hive-2.3

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -203,7 +203,7 @@ parquet-format/2.4.0//parquet-format-2.4.0.jar
203203
parquet-hadoop/1.10.1//parquet-hadoop-1.10.1.jar
204204
parquet-jackson/1.10.1//parquet-jackson-1.10.1.jar
205205
protobuf-java/2.5.0//protobuf-java-2.5.0.jar
206-
py4j/0.10.9//py4j-0.10.9.jar
206+
py4j/0.10.9.7//py4j-0.10.9.7.jar
207207
pyrolite/4.30//pyrolite-4.30.jar
208208
re2j/1.1//re2j-1.1.jar
209209
scala-collection-compat_2.12/2.1.1//scala-collection-compat_2.12-2.1.1.jar

python/docs/Makefile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ SPHINXBUILD ?= sphinx-build
77
PAPER ?=
88
BUILDDIR ?= _build
99

10-
export PYTHONPATH=$(realpath ..):$(realpath ../lib/py4j-0.10.9-src.zip)
10+
export PYTHONPATH=$(realpath ..):$(realpath ../lib/py4j-0.10.9.7-src.zip)
1111

1212
# User-friendly check for sphinx-build
1313
ifeq ($(shell which $(SPHINXBUILD) >/dev/null 2>&1; echo $$?), 1)

python/docs/make2.bat

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
REM Command file for Sphinx documentation
44

5-
set PYTHONPATH=..;..\lib\py4j-0.10.9-src.zip
5+
set PYTHONPATH=..;..\lib\py4j-0.10.9.7-src.zip
66

77

88
if "%SPHINXBUILD%" == "" (

python/lib/py4j-0.10.9-src.zip

-40.6 KB
Binary file not shown.

python/lib/py4j-0.10.9.7-src.zip

44.8 KB
Binary file not shown.

python/setup.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -207,7 +207,7 @@ def _supports_symlinks():
207207
'pyspark.examples.src.main.python': ['*.py', '*/*.py']},
208208
scripts=scripts,
209209
license='http://www.apache.org/licenses/LICENSE-2.0',
210-
install_requires=['py4j==0.10.9'],
210+
install_requires=['py4j==0.10.9.7'],
211211
extras_require={
212212
'ml': ['numpy>=1.7'],
213213
'mllib': ['numpy>=1.7'],

sbin/spark-config.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,6 +28,6 @@ export SPARK_CONF_DIR="${SPARK_CONF_DIR:-"${SPARK_HOME}/conf"}"
2828
# Add the PySpark classes to the PYTHONPATH:
2929
if [ -z "${PYSPARK_PYTHONPATH_SET}" ]; then
3030
export PYTHONPATH="${SPARK_HOME}/python:${PYTHONPATH}"
31-
export PYTHONPATH="${SPARK_HOME}/python/lib/py4j-0.10.9-src.zip:${PYTHONPATH}"
31+
export PYTHONPATH="${SPARK_HOME}/python/lib/py4j-0.10.9.7-src.zip:${PYTHONPATH}"
3232
export PYSPARK_PYTHONPATH_SET=1
3333
fi

0 commit comments

Comments
 (0)