Skip to content

Latest commit

 

History

History

rfa-spark

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 

Risers Fatigue Analysis Synthetic - Spark

How to Run

Dependencies:

Setup and configuration:

Clone repository:

$ git clone https://github.com/vssousa/d-chiron.git
$ cd d-chiron

Edit the input file:

$ vi rfa-synthetic/rfa-spark/input.dataset
  • Example:
ID;SPLITMAP;SPLITFACTOR;MAP1;MAP2;FILTER1;F1;FILTER2;F2;REDUCE;REDUCEFACTOR
1;5;8;5;5;5;50;5;50;5;4
  • Fields:
  • ID: Entry identifier
  • SPLITMAP: Average Task Cost in Uncompress activity (seconds)
  • SPLITFACTOR: Number of entries in the input dataset after uncompression
  • MAP1: Average Task Cost in Pre-Processing activity (seconds)
  • MAP2: Average Task Cost in Analyze Riser sactivity (seconds)
  • FILTER1:Average Task Cost in Calculate Wear and Tear activity (seconds)
  • F1: Amount of entries for Calculate Wear and Tear activity to filter in % (i.e., Percentage that will continue in the flow)
  • FILTER2:Average Task Cost in Analyze Position activity (seconds)
  • F2: Amount of entries for Analyze Position activity to filter in %(i.e., Percentage that will continue in the flow)
  • REDUCE: Average Task Cost in Compress Results activity (seconds)
  • REDUCEFACTOR: Number of compressed output entries

Run

$ export SPARK_HOME=/path/to/spark
  • Change directory to rfa-spark:
$ cd rfa-synthetic/rfa-spark
  • Run:
$ ./run.sh <spark-master-url> <num-executors> <total-executor-cores>

Where:

  • spark-master-url: The master URL for the cluster

  • num-executors: Number of Apache Spark executors requested on thec cluster.

  • total-executor-cores: Total Number of cores requested on the cluster.

  • Example:

$ ./run.sh  spark://hostname:7077 1 2

Source and Build:

Source Code

Build Dependencies

Build

  • Change directory to rfa-spark-project:
$ cd rfa-synthetic/rfa-spark/rfa-spark-project
  • Maven
$ mvn package