A Python module that provides the tools you need to authenticate with, and use the Qubole Data Service API.
Run the following command (may need to do this as root):
$ python setup.py install
This should place a command line utility 'qds.py' somewhere in your path
$ which qds.py
/usr/bin/qds.py
qds.py allows running Hive, Hadoop, Pig and Shell commands against QDS. Users can run commands synchronously - or submit a command and check it's status.
$ qds.py -h # will print detailed usage
Examples:
-
run a hive query and print the results
$ qds.py --token 'xxyyzz' hivecmd run --query "show tables" $ qds.py --token 'xxyyzz' hivecmd run --script_location /tmp/myquery $ qds.py --token 'xxyyzz' hivecmd run --script_location s3://my-qubole-location/myquery
-
pass in api token from bash environment variable
$ export QDS_API_TOKEN=xxyyzz
-
run the example hadoop command
$ qds.py hadoopcmd run streaming -files 's3n://paid-qubole/HadoopAPIExamples/WordCountPython/mapper.py,s3n://paid-qubole/HadoopAPIExamples/WordCountPython/reducer.py' -mapper mapper.py -reducer reducer.py -numReduceTasks 1 -input 's3n://paid-qubole/default-datasets/gutenberg' -output 's3n://example.bucket.com/wcout'
-
check the status of command # 12345678
$ qds.py hivecmd check 12345678 {"status": "done", ... }
An example Python application needs to do the following:
-
Set the api_token:
from qds_sdk.qubole import Qubole Qubole.configure(api_token='ksbdvcwdkjn123423')
-
Use the Command classes defined in commands.py to execute commands. To run Hive Command:
from qds_sdk.commands import * hc=HiveCommand.create(query='show tables') print "Id: %s, Status: %s" % (str(hc.id), hc.status)
example/mr_1.py contains a Hadoop Streaming example