Skip to content

Commit c7c03f1

Browse files
LuciferYangcloud-fan
authored andcommitted
[SPARK-40073][INFRA][BUILD][CORE][SQL][AVRO][PYTHON] Replace external/{moduleName} with connector/{moduleName}
### What changes were proposed in this pull request? SPARK-38569 rename `external` top level dir to `connector`, but the use of `external/${moduleName}` still remains in Spark, this pr replaces them all with `connector/{moduleName}` ### Why are the changes needed? Should use `connector/{moduleName}` instead of `external/${moduleName} ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? Pass GitHub Actions Closes apache#37512 from LuciferYang/external-2-connector. Authored-by: yangjie01 <[email protected]> Signed-off-by: Wenchen Fan <[email protected]>
1 parent 87094f8 commit c7c03f1

File tree

8 files changed

+18
-18
lines changed

8 files changed

+18
-18
lines changed

.github/labeler.yml

+2-2
Original file line numberDiff line numberDiff line change
@@ -103,7 +103,7 @@ SQL:
103103
- "**/*schema.R"
104104
- "**/*types.R"
105105
AVRO:
106-
- "external/avro/**/*"
106+
- "connector/avro/**/*"
107107
- "python/pyspark/sql/avro/**/*"
108108
DSTREAM:
109109
- "streaming/**/*"
@@ -123,7 +123,7 @@ MLLIB:
123123
- "python/pyspark/mllib/**/*"
124124
STRUCTURED STREAMING:
125125
- "**/sql/**/streaming/**/*"
126-
- "external/kafka-0-10-sql/**/*"
126+
- "connector/kafka-0-10-sql/**/*"
127127
- "python/pyspark/sql/streaming/**/*"
128128
- "python/pyspark/sql/tests/test_streaming.py"
129129
- "**/*streaming.R"

LICENSE

+1-1
Original file line numberDiff line numberDiff line change
@@ -216,7 +216,7 @@ core/src/main/resources/org/apache/spark/ui/static/bootstrap*
216216
core/src/main/resources/org/apache/spark/ui/static/jsonFormatter*
217217
core/src/main/resources/org/apache/spark/ui/static/vis*
218218
docs/js/vendor/bootstrap.js
219-
external/spark-ganglia-lgpl/src/main/java/com/codahale/metrics/ganglia/GangliaReporter.java
219+
connector/spark-ganglia-lgpl/src/main/java/com/codahale/metrics/ganglia/GangliaReporter.java
220220

221221

222222
Python Software Foundation License

connector/avro/src/test/scala/org/apache/spark/sql/avro/AvroSuite.scala

+1-1
Original file line numberDiff line numberDiff line change
@@ -1816,7 +1816,7 @@ abstract class AvroSuite
18161816
// It generates input files for the test below:
18171817
// "SPARK-31183, SPARK-37705: compatibility with Spark 2.4/3.2 in reading dates/timestamps"
18181818
ignore("SPARK-31855: generate test files for checking compatibility with Spark 2.4/3.2") {
1819-
val resourceDir = "external/avro/src/test/resources"
1819+
val resourceDir = "connector/avro/src/test/resources"
18201820
val version = SPARK_VERSION_SHORT.replaceAll("\\.", "_")
18211821
def save(
18221822
in: Seq[String],

connector/kinesis-asl/src/main/python/examples/streaming/kinesis_wordcount_asl.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -36,8 +36,8 @@
3636
3737
# run the example
3838
$ bin/spark-submit --jars \
39-
'external/kinesis-asl-assembly/target/spark-streaming-kinesis-asl-assembly_*.jar' \
40-
external/kinesis-asl/src/main/python/examples/streaming/kinesis_wordcount_asl.py \
39+
'connector/kinesis-asl-assembly/target/spark-streaming-kinesis-asl-assembly_*.jar' \
40+
connector/kinesis-asl/src/main/python/examples/streaming/kinesis_wordcount_asl.py \
4141
myAppName mySparkStream https://kinesis.us-east-1.amazonaws.com us-east-1
4242
4343
There is a companion helper class called KinesisWordProducerASL which puts dummy data

dev/checkstyle-suppressions.xml

+1-1
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@
3131
<suppress checks=".*"
3232
files="core/src/main/java/org/apache/spark/util/collection/TimSort.java"/>
3333
<suppress checks=".*"
34-
files="external/spark-ganglia-lgpl/src/main/java/com/codahale/metrics/ganglia/GangliaReporter.java"/>
34+
files="connector/spark-ganglia-lgpl/src/main/java/com/codahale/metrics/ganglia/GangliaReporter.java"/>
3535
<suppress checks=".*"
3636
files="sql/core/src/main/java/org/apache/spark/sql/api.java/*"/>
3737
<suppress checks="LineLength"

dev/sparktestsupport/modules.py

+9-9
Original file line numberDiff line numberDiff line change
@@ -253,7 +253,7 @@ def __hash__(self):
253253
name="avro",
254254
dependencies=[sql],
255255
source_file_regexes=[
256-
"external/avro",
256+
"connector/avro",
257257
],
258258
sbt_test_goals=[
259259
"avro/test",
@@ -264,7 +264,7 @@ def __hash__(self):
264264
name="sql-kafka-0-10",
265265
dependencies=[sql],
266266
source_file_regexes=[
267-
"external/kafka-0-10-sql",
267+
"connector/kafka-0-10-sql",
268268
],
269269
sbt_test_goals=[
270270
"sql-kafka-0-10/test",
@@ -309,8 +309,8 @@ def __hash__(self):
309309
name="streaming-kinesis-asl",
310310
dependencies=[tags, core],
311311
source_file_regexes=[
312-
"external/kinesis-asl/",
313-
"external/kinesis-asl-assembly/",
312+
"connector/kinesis-asl/",
313+
"connector/kinesis-asl-assembly/",
314314
],
315315
build_profile_flags=[
316316
"-Pkinesis-asl",
@@ -327,9 +327,9 @@ def __hash__(self):
327327
dependencies=[streaming, core],
328328
source_file_regexes=[
329329
# The ending "/" is necessary otherwise it will include "sql-kafka" codes
330-
"external/kafka-0-10/",
331-
"external/kafka-0-10-assembly",
332-
"external/kafka-0-10-token-provider",
330+
"connector/kafka-0-10/",
331+
"connector/kafka-0-10-assembly",
332+
"connector/kafka-0-10-token-provider",
333333
],
334334
sbt_test_goals=["streaming-kafka-0-10/test", "token-provider-kafka-0-10/test"],
335335
)
@@ -761,15 +761,15 @@ def __hash__(self):
761761
dependencies=[],
762762
build_profile_flags=["-Pspark-ganglia-lgpl"],
763763
source_file_regexes=[
764-
"external/spark-ganglia-lgpl",
764+
"connector/spark-ganglia-lgpl",
765765
],
766766
)
767767

768768
docker_integration_tests = Module(
769769
name="docker-integration-tests",
770770
dependencies=[sql],
771771
build_profile_flags=["-Pdocker-integration-tests"],
772-
source_file_regexes=["external/docker-integration-tests"],
772+
source_file_regexes=["connector/docker-integration-tests"],
773773
sbt_test_goals=["docker-integration-tests/test"],
774774
environ=None
775775
if "GITHUB_ACTIONS" not in os.environ

python/pyspark/sql/avro/functions.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -140,7 +140,7 @@ def _test() -> None:
140140
import sys
141141
from pyspark.testing.utils import search_jar
142142

143-
avro_jar = search_jar("external/avro", "spark-avro", "spark-avro")
143+
avro_jar = search_jar("connector/avro", "spark-avro", "spark-avro")
144144
if avro_jar is None:
145145
print(
146146
"Skipping all Avro Python tests as the optional Avro project was "

python/pyspark/testing/streamingutils.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@
3535
)
3636
else:
3737
kinesis_asl_assembly_jar = search_jar(
38-
"external/kinesis-asl-assembly",
38+
"connector/kinesis-asl-assembly",
3939
"spark-streaming-kinesis-asl-assembly-",
4040
"spark-streaming-kinesis-asl-assembly_",
4141
)

0 commit comments

Comments
 (0)