Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to make Flint work with Spark 3.0 #83

Open
yitao-li opened this issue May 22, 2020 · 3 comments
Open

How to make Flint work with Spark 3.0 #83

yitao-li opened this issue May 22, 2020 · 3 comments

Comments

@yitao-li
Copy link

yitao-li commented May 22, 2020

Hi n00b question: given how awesome and popular Flint has been, I'm really interested in making it work with Spark 3.0.

So I went ahead and tried the changes in https://github.com/twosigma/flint/pull/82/files which made Flint build successfully with Spark 3.0-preview2, but now some tests are failing (see
flint-spark-3.0.0-build.log ), and I'm not completely sure how to fix the test failures or how much work it might take to fix them.

Any idea how to resolve the test failures? Thanks in advance!

@yitao-li yitao-li changed the title How to make Flint with Spark 3.0 How to make Flint work with Spark 3.0 May 22, 2020
@yitao-li
Copy link
Author

Update: just realized most (if not all) of the test failures are due to org.apache.spark.sql.execution.ColumnarToRowExec not being considered in isPartitionPreservingUnaryNode. I think same for RowToColumnarExec as well.

@dgrnbrg
Copy link

dgrnbrg commented Nov 5, 2020

I'm interesting in this as well. Please post other things you find.

@yitao-li
Copy link
Author

yitao-li commented Nov 5, 2020

@dgrnbrg Hey at the moment I have a branch named 'flint-spark-3.0' in my unofficial fork of Flint that has all necessary changes I'm aware of that will make Flint work with Spark 3.0.

You can view the source code in https://github.com/yitao-li/flint/tree/flint-spark-3.0 and see commit history to inspect changes I made on top of the original Flint source code.

As far as I know all the summarizer functionalities and ASOF join functionalities are working as expected with Spark 3.0 after my changes (in other words, it looks like all changes I have made so far were just to make compiler happy when building with Spark 3.0 and there was no functional change).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants