-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
java.lang.IllegalArgumentException when using parquet file #69
Comments
But if I run the Main.scala in intelliJ on my base windows machine, it executes with no problem.
Note: I hardcoded the config file in ConfigParser.scala
And also, I hardcoded Spark to run locally in the Main.scala like:
Environment Details of base machine:
sbt version:
|
What happens when you set the numKeyCols: 2
numErrorsToReport: 742
tables:
- parquetFile: /home/jyoti/Spark/userdata1.parquet
checks:
- type: nullCheck
column: salary
threshold: 0
# or
threshold: "0" It should be optional, though. We've almost always specified it. |
Actually, I found it: DV doesn't support Spark 3 yet, so all bets are off. But try something: apply this patch to change the literals:
|
We may support Spark 3 after #84. |
Thanks Colin, I will check it out... 😃 |
When trying to run a config check on a parquet file, the following error can be seen:
Ran a spark-submit job as follows:
The config.yaml file has the following content:
I got the userdata1.parquet from the following github link:
https://github.com/Teradata/kylo/blob/master/samples/sample-data/parquet/userdata1.parquet
Environment Details:
latest source code: data-validator-0.13.0
Lubuntu 18.04 LTS x64 version on VMWare Player
4 CPU cores and 2GB ram
Java version
lsb_release output:
uname -s:
sbt -version:
Please let me know if you need anything else.
The text was updated successfully, but these errors were encountered: