Skip to content

A implementation in Scala of CF, Content Based, Sequential and hybrid recommender systems for Spark

License

Notifications You must be signed in to change notification settings

danitico/recsys-spark

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

42 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

recsys-spark

A implementation in Scala of CF, Content Based, Sequential and hybrid recommender systems for Spark.

How can you install the dependencies of this project?

Java & Scala

We will use sdkman to install java and scala.

curl -s "https://get.sdkman.io" | bash
source "$HOME/.sdkman/bin/sdkman-init.sh"

Then, we will install java and scala.

sdk install scala 2.12.15
sdk install java 18.0.2-open

How can you check if it was installed correctly?

Run the following command:

java -version
scala -version

Apache Spark and Apache Hadoop

Run the following commands:

wget https://archive.apache.org/dist/spark/spark-3.2.1/spark-3.2.1-bin-hadoop3.2.tgz
tar -xvf spark-3.2.1-bin-hadoop3.2.tgz
mv spark-3.2.1-bin-hadoop3.2 /opt

After that, add the following lines to .bashrc or .zshrc

export SPARK_HOME=/opt/spark-3.2.1-bin-hadoop3.2
export HADOOP_HOME=/opt/spark-3.2.1-bin-hadoop3.2

And apply those changes in your configuration running:

source $HOME/<your_configuration_file>

How can you check if it was installed correctly?

Run the following command:

spark-shell

How can you run this project?

First compile the project:

sbt compile

After that, you can run the Main.scala file with;

sbt run