Skip to content

Commit 3e4d8e2

Browse files
authored
Merge branch 'apache:master' into master
2 parents 52a77d1 + 3d0d54d commit 3e4d8e2

File tree

514 files changed

+16094
-28318
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

514 files changed

+16094
-28318
lines changed

.gitattributes

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,4 @@
1+
* text=auto eol=lf
2+
*.bat text eol=crlf
3+
*.cmd text eol=crlf
4+
*.png binary

.github/linters/.markdown-lint.yml

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -30,17 +30,11 @@ MD034: false
3030
# no-emphasis-as-heading - Emphasis used instead of a heading
3131
MD036: false
3232

33-
# no-space-in-code - Spaces inside code span elements
34-
MD038: false
35-
3633
# fenced-code-language - Fenced code blocks should have a language specified
3734
MD040: false
3835

3936
# first-line-heading/first-line-h1 - First line in a file should be a top-level heading
4037
MD041: false
4138

42-
# no-alt-text - Images should have alternate text (alt text)
43-
MD045: false
44-
4539
# code-block-style - Code block style
4640
MD046: false

.github/linters/.yaml-lint.yml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,8 +4,6 @@ extends: default
44

55
rules:
66
colons: disable
7-
comments: disable
8-
comments-indentation: disable
97
document-start: disable
108
line-length: disable
119
truthy: false

.github/linters/ruff.toml

Lines changed: 0 additions & 77 deletions
This file was deleted.

.github/pull_request_template.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,9 @@
1010

1111
- Yes, the URL of the associated JIRA ticket is https://issues.apache.org/jira/browse/SEDONA-XXX. The PR name follows the format `[SEDONA-XXX] my subject`.
1212

13-
- No, this is a documentation update. The PR name follows the format `[DOCS] my subject`.
13+
- No:
14+
- this is a documentation update. The PR name follows the format `[DOCS] my subject`
15+
- this is a CI update. The PR name follows the format `[CI] my subject`
1416

1517

1618
## What changes were proposed in this PR?

.github/workflows/docker-build.yml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ jobs:
2424
strategy:
2525
fail-fast: true
2626
matrix:
27-
os: ['ubuntu-latest', 'macos-12']
27+
os: ['ubuntu-latest']
2828
spark: ['3.5.1', '3.4.3', '3.3.4']
2929
include:
3030
- spark: 3.5.1
@@ -55,6 +55,7 @@ jobs:
5555
- name: Setup docker (missing on macOS)
5656
if: runner.os == 'macos'
5757
run: |
58+
brew install colima
5859
brew install docker
5960
colima start
6061
DOCKER_CONFIG=${DOCKER_CONFIG:-$HOME/.docker}

.github/workflows/java.yml

Lines changed: 0 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -68,18 +68,6 @@ jobs:
6868
scala: 2.12.15
6969
jdk: '8'
7070
skipTests: ''
71-
- spark: 3.2.3
72-
scala: 2.12.15
73-
jdk: '8'
74-
skipTests: ''
75-
- spark: 3.1.2
76-
scala: 2.12.15
77-
jdk: '8'
78-
skipTests: ''
79-
- spark: 3.0.3
80-
scala: 2.12.15
81-
jdk: '8'
82-
skipTests: ''
8371
steps:
8472
- uses: actions/checkout@v4
8573
- uses: actions/setup-java@v4

.github/workflows/python.yml

Lines changed: 36 additions & 40 deletions
Original file line numberDiff line numberDiff line change
@@ -41,57 +41,37 @@ jobs:
4141
- spark: '3.5.0'
4242
scala: '2.12.8'
4343
python: '3.10'
44-
hadoop: '3'
4544
shapely: '1'
4645
- spark: '3.5.0'
4746
scala: '2.12.8'
4847
python: '3.10'
49-
hadoop: '3'
5048
- spark: '3.5.0'
5149
scala: '2.12.8'
5250
python: '3.9'
53-
hadoop: '3'
5451
- spark: '3.5.0'
5552
scala: '2.12.8'
5653
python: '3.8'
57-
hadoop: '3'
5854
- spark: '3.4.0'
5955
scala: '2.12.8'
6056
python: '3.10'
61-
hadoop: '3'
6257
- spark: '3.4.0'
6358
scala: '2.12.8'
6459
python: '3.9'
65-
hadoop: '3'
6660
- spark: '3.4.0'
6761
scala: '2.12.8'
6862
python: '3.8'
69-
hadoop: '3'
7063
- spark: '3.4.0'
7164
scala: '2.12.8'
7265
python: '3.7'
73-
hadoop: '3'
7466
- spark: '3.4.0'
7567
scala: '2.12.8'
7668
python: '3.7'
77-
hadoop: '3'
7869
shapely: '1'
7970
- spark: '3.3.0'
8071
scala: '2.12.8'
8172
python: '3.8'
82-
hadoop: '3'
83-
- spark: '3.2.0'
84-
scala: '2.12.8'
85-
python: '3.7'
86-
hadoop: '2.7'
87-
- spark: '3.1.2'
88-
scala: '2.12.8'
89-
python: '3.7'
90-
hadoop: '2.7'
91-
- spark: '3.0.3'
92-
scala: '2.12.8'
93-
python: '3.7'
94-
hadoop: '2.7'
73+
env:
74+
VENV_PATH: /home/runner/.local/share/virtualenvs/python-${{ matrix.python }}
9575
steps:
9676
- uses: actions/checkout@v4
9777
- uses: actions/setup-java@v4
@@ -113,18 +93,6 @@ jobs:
11393
run: |
11494
SPARK_COMPAT_VERSION=${SPARK_VERSION:0:3}
11595
mvn -q clean install -DskipTests -Dspark=${SPARK_COMPAT_VERSION} -Dscala=${SCALA_VERSION:0:4} -Dgeotools
116-
- env:
117-
SPARK_VERSION: ${{ matrix.spark }}
118-
HADOOP_VERSION: ${{ matrix.hadoop }}
119-
run: |
120-
wget https://archive.apache.org/dist/spark/spark-${SPARK_VERSION}/spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz
121-
wget https://repo.osgeo.org/repository/release/javax/media/jai_core/${JAI_CORE_VERSION}/jai_core-${JAI_CORE_VERSION}.jar
122-
wget https://repo.osgeo.org/repository/release/javax/media/jai_codec/${JAI_CODEC_VERSION}/jai_codec-${JAI_CODEC_VERSION}.jar
123-
wget https://repo.osgeo.org/repository/release/javax/media/jai_imageio/${JAI_IMAGEIO_VERSION}/jai_imageio-${JAI_IMAGEIO_VERSION}.jar
124-
tar -xzf spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz
125-
mv -v jai_core-${JAI_CORE_VERSION}.jar spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}/jars/
126-
mv -v jai_codec-${JAI_CODEC_VERSION}.jar spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}/jars/
127-
mv -v jai_imageio-${JAI_IMAGEIO_VERSION}.jar spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}/jars/
12896
- run: sudo apt-get -y install python3-pip python-dev-is-python3
12997
- run: sudo pip3 install -U setuptools
13098
- run: sudo pip3 install -U wheel
@@ -141,15 +109,43 @@ jobs:
141109
echo "Patching Pipfile to use Shapely 1.x"
142110
sed -i 's/^shapely.*$/shapely="<2.0.0"/g' Pipfile
143111
fi
112+
export PIPENV_CUSTOM_VENV_NAME=python-${PYTHON_VERSION}
144113
pipenv --python ${PYTHON_VERSION}
145114
pipenv install pyspark==${SPARK_VERSION}
146115
pipenv install --dev
147116
pipenv graph
148117
- env:
149-
SPARK_VERSION: ${{ matrix.spark }}
150-
HADOOP_VERSION: ${{ matrix.hadoop }}
151-
run: find spark-shaded/target -name sedona-*.jar -exec cp {} spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}/jars/ \;
118+
PYTHON_VERSION: ${{ matrix.python }}
119+
run: |
120+
wget --retry-connrefused --waitretry=10 --read-timeout=20 --timeout=15 --tries=5 https://repo.osgeo.org/repository/release/javax/media/jai_core/${JAI_CORE_VERSION}/jai_core-${JAI_CORE_VERSION}.jar
121+
wget --retry-connrefused --waitretry=10 --read-timeout=20 --timeout=15 --tries=5 https://repo.osgeo.org/repository/release/javax/media/jai_codec/${JAI_CODEC_VERSION}/jai_codec-${JAI_CODEC_VERSION}.jar
122+
wget --retry-connrefused --waitretry=10 --read-timeout=20 --timeout=15 --tries=5 https://repo.osgeo.org/repository/release/javax/media/jai_imageio/${JAI_IMAGEIO_VERSION}/jai_imageio-${JAI_IMAGEIO_VERSION}.jar
123+
mv -v jai_core-${JAI_CORE_VERSION}.jar ${VENV_PATH}/lib/python${PYTHON_VERSION}/site-packages/pyspark/jars
124+
mv -v jai_codec-${JAI_CODEC_VERSION}.jar ${VENV_PATH}/lib/python${PYTHON_VERSION}/site-packages/pyspark/jars
125+
mv -v jai_imageio-${JAI_IMAGEIO_VERSION}.jar ${VENV_PATH}/lib/python${PYTHON_VERSION}/site-packages/pyspark/jars
152126
- env:
153-
SPARK_VERSION: ${{ matrix.spark }}
154-
HADOOP_VERSION: ${{ matrix.hadoop }}
155-
run: (export SPARK_HOME=$PWD/spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION};export PYTHONPATH=$SPARK_HOME/python;cd python;pipenv run pytest tests)
127+
PYTHON_VERSION: ${{ matrix.python }}
128+
run: find spark-shaded/target -name sedona-*.jar -exec cp {} ${VENV_PATH}/lib/python${PYTHON_VERSION}/site-packages/pyspark/jars/ \;
129+
- env:
130+
PYTHON_VERSION: ${{ matrix.python }}
131+
run: |
132+
export SPARK_HOME=${VENV_PATH}/lib/python${PYTHON_VERSION}/site-packages/pyspark
133+
cd python
134+
source ${VENV_PATH}/bin/activate
135+
pytest tests
136+
- env:
137+
PYTHON_VERSION: ${{ matrix.python }}
138+
run: |
139+
if [ ! -f "${VENV_PATH}/lib/python${PYTHON_VERSION}/site-packages/pyspark/sbin/start-connect-server.sh" ]
140+
then
141+
echo "Skipping connect tests for Spark $SPARK_VERSION"
142+
exit
143+
fi
144+
145+
export SPARK_HOME=${VENV_PATH}/lib/python${PYTHON_VERSION}/site-packages/pyspark
146+
export SPARK_REMOTE=local
147+
148+
cd python
149+
source ${VENV_PATH}/bin/activate
150+
pip install "pyspark[connect]==${SPARK_VERSION}"
151+
pytest tests/sql/test_dataframe_api.py

.github/workflows/r.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@ jobs:
3232
strategy:
3333
fail-fast: true
3434
matrix:
35-
spark: [3.0.3, 3.1.2, 3.2.1, 3.3.0, 3.4.0, 3.5.0]
35+
spark: [3.3.0, 3.4.0, 3.5.0]
3636
hadoop: [3]
3737
scala: [2.12.15]
3838
r: [oldrel, release]

0 commit comments

Comments
 (0)