LERS algorithm in Hadoop MapReduce Download source code here: http://webpages.uncc.edu/aatzache/ITCS6190/Exercises/SparkLERS_Example.zip To run the Code in AWS EMR cluster: 1. Create a cluster with Hadoop and Spark in AWS and start the cluster. Once the cluster is running, log-in to the master node using Putty(Windows) or SSH(MAC or Linux). Refer to: GroupActivity_01 Instructions for logging in to the AWS EMR cluster Simple Commands Task2 http://webpages.uncc.edu/aatzache/ITCS6190/Exercises/GroupActivity01_LoggingInto_AWS_Cluster_SimpleCommands_Task2.doc video: AWS-EMR_Cluster_Setup https://youtu.be/_A_xEtd2OIM 2. Create a data bucket in AWS S3. Upload data.txt, attributes.txt, parameters.txt from the Data directory and SparkLERS.jar files to S3 3. From the master node download SparkLERS.jar using the command: aws s3 cp s3://BUCKET_NAME/SparkLERS.jar . 4. Run the .jar file using your terminal or Putty using following command: spark-submit --class "org.ActionRules.Main" --master yarn --deploy-mode client LOCATION_OF_JAR_FILE s3://BUCKET_NAME/attributes.txt s3://BUCKET_NAME/parameters.txt s3://BUCKET_NAME/data.txt s3://BUCKET_NAME/SparkLERSOutput 5. Download the output folder(SparkLERSOutput) from S3 to your local machine 6. Submit the program output in Canvas 7. Save all commands in your Terminal Window in a Text File , including your login and username until the last command, with all commands, results and errors if any from the terminal window and upload the TeminalWindow.txt to Canvas . 8. Delete/Terminate the AWS cluster and delete all files from S3 when finished, otherwise Amazon will charge your Credit Card