Skip to content

A Scala / Java / Python library for cleansing, transforming and preparing large datasets for ML operations on Apache Spark.

License

Notifications You must be signed in to change notification settings

data-commons/prep-buddy

Repository files navigation

Prep Buddy

Build Status

Data Preparation Library for Spark

A Scala / Java / Python library for cleaning, transforming and executing other preparation tasks for large datasets on Apache Spark.

It is currently maintained by a team of developers from ThoughtWorks.

Post questions and comments to the Google group, or email them directly to [email protected]

Docs are available at http://data-commons.github.io/prep-buddy Or check out the Scaladocs.

Our aim is to provide a set of algorithms for cleaning and transforming very large data sets,
inspired by predecessors such as Open Refine, Pandas and Scikit-learn packages.

Important links

Usage!

To use this library, add a maven dependency to datacommons in your project:

<dependency>
    <groupId>com.thoughtworks.datacommons</groupId>
    <artifactId>prep-buddy</artifactId>
    <version>0.5.1</version>
</dependency>

For other build tools check on Maven Repositry

##Python

If you don't have pip. Intsall pip.

pip install prep-buddy

For using pyspark on command-line Download the Jar.

pyspark --jars [PATH-TO-JAR]
spark-submit --driver-class-path [PATH-TO-JAR] [Your python file.]

This library is currently built for Spark 1.6.x, but is also compatible with 1.4.x.

Dependencies

The library depends on a few other libraries.

  • Apache Commons Math for general math and statistics functionality.
  • Apache Spark for all the distributed computation capabilities.
  • Open CSV for parsing the files.

Download

Documentation Wiki

Contributing

  • Create a pull request.

About

A Scala / Java / Python library for cleansing, transforming and preparing large datasets for ML operations on Apache Spark.

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Contributors 5