-
Notifications
You must be signed in to change notification settings - Fork 183
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
"joinLeft" is not working when joining a clock with a parque-based DataFrame #57
Comments
Update: so |
Some googling around might indicate that there's some incompatibility between the framework and the library versions, but your front page does say that the library works with Spark 2.3., and the lib's version (per python) is 0.6.0 |
The error trace is calling this line surely: |
Are you using the databricks jar? The apache jar doesn't work on databricks runtime for some reason. |
Li Jin, I appreciate your reply. I downloaded the flint jar file that is currently loaded onto DataBricks, its size is 2171677 and here is its manifest file:
As I stated previously: Please, assist further! |
Built from source ( |
There is issues with using the standard flint jar on Databricks platform so they build a specific jar for Flint 0.6 for their platform here: https://github.com/databricks/databricks-accelerators/tree/master/projects/databricks-flint |
I downloaded the one from DataBricks, its size is 2252415 and its name is flint_0_6_0_databricks.jar |
@LeoDashTM Could you please try this in the local pyspark notebook with --packages? |
I'm not too familiar yet with local pyspark, but see the 2nd comment of the issue I opened with DataBricks: databricks/databricks-accelerators#1 Ok, actually just tried: The data file is attached to the first comment and both have the reproducible code (the first one for DB and the second one for a local installation). Are you able to run either one of them against Spark 2.3.1 and see the issue? Thanks. |
Update. |
It sounds like it’s a Databricks specific issue. I think the way they
handle third party packages is different from the open source one. That’s
why the Databricks jar exists in the first place.
…On Thu, Nov 1, 2018 at 2:40 AM Leo Dashevskiy ***@***.***> wrote:
Update.
Just installed Spark 2.3.1 locally and running pyspark --packages
com.twosigma:flint:0.6.0 then executing each line by hand seems to work
correctly.
So the deal is indeed with the flavor of Spark installed on DataBricks
then? Is that the right conclusion?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#57 (comment)>, or mute
the thread
<https://github.com/notifications/unsubscribe-auth/AAwbrMP_0bfk5tCTEXdgZH8UIG9QNYIpks5uqpd3gaJpZM4X812S>
.
|
With the Databricks' latest commit: databricks/databricks-accelerators@d23782a The results of a clock creation and printing are a bit unexpected for me:
Are these results as expecte0d @icexelloss ? Are you getting the exact same thing on your setup? And if this is as expected, then how do I create a clock with the range I specified via |
Ok, I tried with TwoSigma's official latest build and the data frame is displayed as expected, so it's a DataBricks issue again... |
Same issue: #58 ? |
@icexelloss
(even though the "timeColumn" argument error can be bypassed by renaming the column in question to
time
) thejoinLeft
is not working for me:With the output being:
By the way, what is the way to even display the contents of the clocks DataFrame?
The second to last commented out line (with the
.show
command) errors out, so I don't understand howTimeSeriesDataFrame
is inheriting from a regularDataFrame
, for which that method is available...The
display
method also fails...Anyways, what is wrong with the
leftJoin
here? The clock is on the left like you indicated it should be. Swapping left and right data frames also does not help.Is this reproducible for you?
Please, advise, if I'm not using/calling it correctly or if it's a bug.
The flint libraries (the Scala and the Python ones) I installed on DataBricks via its UI (from the respective online repos, which might be dated) - I can try and install the latest builds from the freshest source code, if you think that will help.
Thanks.
The text was updated successfully, but these errors were encountered: