Skip to content

nheumann/ddm-spark

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

39 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Setup instructions

This is a basic scala project that includes spark dependencies and necessary build-configuration to build a jar that can be submitted to a spark installation via spark-submit.

Checkout the project

git clone [email protected]:HPI-Information-Systems/spark-tutorial.git

Install IntelliJ

  • Select the Scala and sbt plugins to be installed during the installation; otherwise you need to install them later

Import the project into IntelliJ

  • Start Intellij
  • Select Open or Import
  • Select the spark-tutorial/build.sbt file and click OK
  • Click Open as Project (IntelliJ should automatically recognize the sbt project nature; note that the initialization might take a while for missing downloads and indexing; if IntelliJ does not automatically download the depencencies, you can open the sbt tab on the right edge of the screen and click Reload all sbt projects)

Load the homework data

Run the tutorial

  • Switch to your IntelliJ window and open spark-tutorial/src/main/scala/de/hpi/spark_tutorial in the project tree view
  • Right-click SimpleSpark.scala and click Run SimpleSpark

Solve the DDM homework

  • Remove the Tutorial and LongestCommonSubstring calls
  • Implement the Sindy algorithm

Build your final jar file

  • Click Run->Edit Configurations->+->SBT Task
  • Enter a name, such as SparkTutorial assembly
  • Enter task clean assembly
  • Click OK
  • You can now switch to and run the SparkTutorial assembly target in the top right Intellij bar
  • Find your fat-jar in spark-tutorial/target/scala-2.12/SparkTutorialSBT-assembly-0.1.jar

Some further notes for cluster submits

  • Modify the settings in the build.sbt to match the spark installation (Relevant settings are both the scalaVersion parameter and the specific versions of all the spark packages (for example in libraryDependencies += org.apache.spark %% spark-core % x replace x with the version of the spark installation, where you want to execute the code). If the scala version and the spark version do not exactly match those of the spark installation, you will encounter errors while executing the jar that are not really helpful in determining the cause.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages