Skip to content

Compare performance of Apache MapReduce and Spark through SQL queries on Large Dataset

License

Notifications You must be signed in to change notification settings

vishnuys/cloudcomparer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

38 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Cloud Project: Compare performance of Hadoop vs Spark

Rest API to run SQL query on large set of data using Hadoop and Spark.

Requirements:

  • Python 3
  • Hadoop
  • Hadoop Streamer Path
  • HDFS CSV Files Path

Installation:

  1. Create a virtualenv Cloud and activate it. If you do not have virtualenv installed, install it. (Installation: Windows, Linux & MAC OS)
  2. In the project home folder, Install the necessary packages using the command pip install -r requirements .
  3. Create config.py in cloudproject folder.
  4. Add the following code to config.py
    HADOOP_STREAMER_PATH = "<Path of the Hadoop Streamer JAR file>"
    HDFS_CSV_PATH = "<Path of the CSV Files in HDFS>"
  5. Execute the following command to apply migrations: python manage.py migrate.
  6. To run the server in localhost with default port, execute the command python manage.py runserver.

Usage:

This REST API doesn't have any authentication. To access this API, send a POST request using any request managers like Postman with data having key query with value as the SQL query.

The output of the query contains two keys: mapreduce and spark. Each key contains the time it took to execute the query on the specified data and the result of the query.

You can learn more about APIs and how to use them here.

About

Compare performance of Apache MapReduce and Spark through SQL queries on Large Dataset

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •  

Languages