Skip to content

A Hadoop Map-Reduce job to process the DBLP dataset to produce a graph depicting which professors at the CS department of UIC have co-authored publications.

License

Notifications You must be signed in to change notification settings

mayankrastogi/faculty-collaboration

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

34 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CS 441 – Engineering Distributed Objects for Cloud Computing

Homework 2 – Faculty Collaboration


Overview

The objective of this homework was to process the DBLP dataset using Hadoop Map-Reduce framework to find out the number of publications made by faculty in the Department of Computer Science at the University of Illinois at Chicago, and to generate a graph to visualize this output.

Results

The results obtained after running the map reduce job were visualized using Graphviz:

Publications by Faculty of Computer Science - University of Illinois at Chicago

Each node is labelled with the faculty name and their total number of publications. The size and color of the nodes is relative to the log of number of publications. Logarithmic scale was chosen so that the node size of faculty with large number of publications is not too big in comparison to that of faculty with small number of publications. The log of maximum number of publications by any faculty is divided into 10 buckets to assign the colors using set310 color scheme from the Brewer color schemes.

Each edge in the graph is labelled with the number of publications that were co-authored by the faculty connected by those edges. The thickness of the edges is relative to the number of publications denoted by the edge.

Instructions

My Environment

The project was developed using the following environment:

Prerequisites

  • HDP Sandbox set up and deployed on (VMware or VirtualBox). Read this guide for instructions on how to set up and use HDP Sandbox
  • Ability to use SSH and SCP on your system
  • SBT installed on your system
  • dblp.xml downloaded on your system

Running the map reduce job

  1. Clone or download this repository onto your system

  2. Open the Command Prompt (if using Windows) or the Terminal (if using Linux/Mac) and browse to the project directory

  3. Build the project and generate the jar file using SBT

    sbt clean compile assembly
    
  4. Start HDP sandbox VM

  5. Copy the jar file to HDP Sandbox VM

    scp -P 2222 target/scala-2.12/mayank_k_rastogi_hw2-assembly-0.1.jar [email protected]:~/
    
  6. Copy dblp.xml to HDP Sandbox

    scp -P 2222 /path/to/dblp.xml [email protected]:~/
    
  7. SSH into HDP Sandbox

    ssh -p 2222 [email protected]
    
  8. Create input directory on HDFS and copy dblp.xml there

    hdfs dfs -mkdir input_dir
    
    hdfs dfs -put dblp.xml input_dir/
    
  9. Start the map-reduce job

    hadoop jar mayank_k_rastogi_hw2-assembly-0.1.jar input_dir output_dir
    
  10. You may check the logs using the YARN and grep. The application ID will get printed when the map-reduce job is run

    yarn logs -applicationId application_<your-application-id> | grep FacultyCollaboration
    

Generating the .dot file

  1. Once the map-reduce job is finished, copy the output files from output_dir (on HDFS) to map_reduce_output (on HDP Sandbox)

    hdfs dfs -get output_dir map_reduce_output
    
  2. Logout from your SSH session on HDP Sandbox

    exit
    
  3. Copy the map_reduce_output directory to your local system

    scp -P 2222 -r [email protected]:~/map_reduce_output ./
    
  4. Run the GraphGenerator to read map_reduce_output and generate faculty_collaborations.dot file

    sbt "runMain com.mayankrastogi.cs441.hw2.graph.GraphGenerator map_reduce_output"
    

Visualizing the graph using Viz.js

Viz.js is an online tool that uses Graphviz to generate graphs from .dot files. Although, we can use the Graphviz directly on our system, I could not manage to get the desired output from the command-line version of the program. It seemed to ignore parameters such as splines=true and sep=3, resulting in a lot of overlapping of node names in the generated image.

Follow the below instructions to generate an image from the faculty_collaborations.dot file:

  1. Open faculty_collaborations.dot file in a text editor
  2. Copy all the contents of this file
  3. Open Viz.js in your web browser
  4. Paste the contents of faculty_collaborations.dot in the left pane on the website
  5. Choose neato as the engine and the desired output format (say, .png)

Tutorial on how to deploy the map-reduce job on AWS EMR

To see a demo of how this project can be deployed on AWS Elastic Map Reduce (EMR), see this video on YouTube

How to deploy Hadoop Map Reduce Job on AWS Elastic Map Reduce (EMR)

Working of the Map Reduce Job

The dblp.xml file

The dblp.xml file has <dblp> as the root element. Under <dblp>, we can have <article>, <inproceedings>, <proceedings>, <book>, <incollection>, <phdthesis>, <mastersthesis> or <www>. Since <www> holds information about an author and not a publication itself, we are ignoring this tag in our program.

Except <book> and <proceedings>, each of these tags contain one or more <author> tags which denote an author for that publication. The <book> and <proceedings> tags have a similar tag <editor>. We treat both these tags as the same, which means that the presence of a faculty's name in either of these tags will count towards their total number of publications and collaborations.

Sharding the input file into logical splits

The MultiTagXmlInputFormat takes care of sharding the dblp.xml file into logical subsets that are fed into the FacultyCollaborationMapper. It reads the dblp.xml file and looks for one of these start tags – <article , <inproceedings , <proceedings , <book , <incollection , <phdthesis , <mastersthesis . Once a match is found, it stores all the bytes that appear after the matched start tag, into a buffer, until the corresponding end tag is found. This forms our logical split that is then fed into a mapper. The start and end tags to look for can be configured using faculty-collaboration.xml-input.start-tags and faculty-collaboration.xml-input.end-tags configuration settings in application.conf.

Mapping the Input Splits

The FacultyCollaborationMapper takes in a subset of dblp.xml file and parses it using Scala's scala-xml module. The mapper then extracts all the entries within <author> or <editor> tags and matches them against the list of faculty that belong to UIC's CS department, which is defined in src/main/resources/uic-cs-faculty-list.txt. This file maps the different variations of a faculty's name (that are known to appear in the dblp.xml file) to the faculty's name as it appears on the UIC CS Department Website.

The list of faculty found associated with a publication is then sorted in lexicographical order and all combinations of pairs of faculty (nC2 + nC1) are generated from this list, and then joined with --. For example, if the list of authors extracted from the subset is List("Philip S. Yu", "Ouri Wolfson", "Peter Nelson"), then the following combinations will be generated:

Iterator(
    "Ouri Wolfson",
    "Peter Nelson",
    "Philip S. Yu",
    "Ouri Wolfson -- Peter Nelson",
    "Ouri Wolfson -- Philip S. Yu",
    "Peter Nelson -- Philip S. Yu"
)

The sorting step ensures that there is only one way of forming a string that represents the same combination of two faculty.

The mapper then emits these combinations as keys with a value of 1, where each key having one faculty name represents a node and each key having a pair of faculty names represent an edge, such that the text file written to the disk with the results of a mapper will look like this:

Ouri Wolfson	1
Peter Nelson	1
Philip S. Yu	1
Ouri Wolfson -- Peter Nelson	1
Ouri Wolfson -- Philip S. Yu	1
Peter Nelson -- Philip S. Yu	1

Reducing the output from the mappers

The job of FacultyCollaborationReducer is very simple. It simply adds up all the values of the same key to get the final count of number of publications by each individual faculty and each faculty pair. The output from the reducer is in the same format as that of the mapper.

Interpreting the Map-Reduce output to generate a graph

The GraphGenerator reads the output files generated by the map-reduce job and generates a .dot file that can in turn be used to generate a graph using Graphviz.

Since the output from the map-reduce job may consist of multiple part files, the program first concatenates all the files present in the specified input directory, except _SUCCESS. It then splits all the lines by new-line characters (\n, \r). Each line is then split by the tab character (\t) to get the keys and values. Separate lists of nodes and edges is then created depending on whether the keys contain -- or not.

The log of maximum number of publications by any faculty is divided into 10 buckets to assign the colors using set310 color scheme from the Brewer color schemes. The size of each node is determined by log(number of publications) / log(maximum number of publications by any faculty). Similarly, the thickness of the edges is determined by (number of publications co-authored by the faculty pair) / (maximum number of publications co-authored by any faculty pair).

The above information is used to write the output in .dot format.

About

A Hadoop Map-Reduce job to process the DBLP dataset to produce a graph depicting which professors at the CS department of UIC have co-authored publications.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published