Skip to content

HTTPS clone URL

Subversion checkout URL

You can clone with HTTPS or Subversion.

Download ZIP
JavaScript Python
branch: master

Fetching latest commit…

Cannot retrieve the latest commit at this time

Failed to load latest commit information.
apps
docs
static
templates
tools
.gitignore
README
__init__.py
connect_redis.py
manage.py
requirements.txt
settings.py
urls.py

README

Hello!!  Welcome to bub.ly!  Here you can upload files and shorten links. You can even use a (applescript) droplet on your dock to upload files easily. 

The site consists of two main pages: a homepage with a login/sign up form, and an account page that you see once you are logged in. The account page lists all the links you have shortened and all the files you have uploaded. The account page also provides you with an area to upload files, a download link for your personalized applescript droplet, and a draggable bookmarklet to shorten links.

DEPENDENCIES

$ sudo pip install -r requirements.txt

EXTERNAL DEPENDENCIES:

 - redis (http://code.google.com/p/redis/)
   - please use the version 2.1.1 as we use hashes feature
   - download link (http://github.com/antirez/redis/tarball/v2.1.1-watch)
   - installation 
       $ tar -xzf redis-2.1.1.tar.gz
       $ cd redis-2.1.1
       $ make
       $ ./redis-server
   - or use redis-server in the tools folder
       $ ./tools/redis-server tools/redis.conf

   
USING DROPLET TO UPLOAD FILES

For Optional Amazon S3 Hosting for files:
RUNNING pyres message queue:

 - pyres is a message queue built on redis and python. (http://github.com/binarydud/pyres)
   The use of pyres is completely optional for this project. 

 - INSTALLATION
 
    $ sudo pip install boto==1.9b pyres==0.9 itty==0.6.4 pystache==0.2.0 redis==2.0.0
    
    Note:  Installing pyres may revert the redis to 1.34 version. We need
           2.0.0 since we use redis hashes

 - There is no need to run pyres to the project working. But it will used to 
   upload the files to S3 asynchronously to reduce the server costs. Similarly
   there is a job to delete the file from S3 when the entry is killed by user.
 
  $ pyres_worker fileupload filedelete
  
  This starts a worker which polls on fileupload and filedelete queue. The 
  current file upload view needs to place the job into the queue which is 
  currently not implemented.


Something went wrong with that request. Please try again.