Permalink
Browse files

Removed egg-info, updated hgignore.

  • Loading branch information...
1 parent 9cb29c9 commit edda19854cd7f0b95cd9efc4d5865b295d83ac3b @shazow committed Sep 18, 2011
View
@@ -1,8 +1,7 @@
syntax: glob
-*.pyc
.*
-*.pidaproject
-*.db
+*.pyc
+*.egg-info
*.log
dist
build
View
@@ -1,82 +0,0 @@
-Metadata-Version: 1.0
-Name: urllib3
-Version: 0.5
-Summary: HTTP library with thread-safe connection pooling and file post support
-Home-page: http://code.google.com/p/urllib3/
-Author: Andrey Petrov
-Author-email: andrey.petrov@shazow.net
-License: MIT
-Download-URL: http://urllib3.googlecode.com/files/urllib3-0.5.tar.gz
-Description: Highlights
- ==========
-
- * Re-use the same socket connection for multiple requests
- (``HTTPConnectionPool`` and ``HTTPSConnectionPool``) (with client-side certificates)
- * File posting (``encode_multipart_formdata``)
- * Built-in redirection and retries (optional)
- * Supports gzip and deflate (big thanks to niphlod)
- * Thread-safe
- * Small and easy to understand codebase perfect for extending and building upon. For a more comprehensive alternative, have a look at `httplib2 <http://code.google.com/p/httplib2/>`_.
-
- What's wrong with urllib and urllib2?
- =====================================
-
- There are two critical features missing from the Python standard library:
- Connection re-using/pooling and file posting. It's not terribly hard to
- implement these yourself, but it's much easier to use a module that already
- did the work for you.
-
- The Python standard libraries ``urllib`` and ``urllib2`` have little to do
- with each other. They were designed to be independent and standalone, each
- solving a different scope of problems, and ``urllib3`` follows in a similar
- vein.
-
- Why do I want to reuse connections?
- ===================================
-
- Performance. When you normally do a urllib call, a separate socket
- connection is created with each request. By reusing existing sockets
- (supported since HTTP 1.1), the requests will take up less resources on the
- server's end, and also provide a faster response time at the client's end.
- With some simple benchmarks (see `test/benchmark.py
- <http://code.google.com/p/urllib3/source/browse/trunk/test/benchmark.py>`_
- ), downloading 15 URLs from google.com is about twice as fast when using
- HTTPConnectionPool (which uses 1 connection) than using plain urllib (which
- uses 15 connections).
-
- This library is perfect for:
-
- * Talking to an API
- * Crawling a website
- * Any situation where being able to post files, handle redirection, and
- retrying is useful. It's relatively lightweight, so it can be used for
- anything!
-
- Examples
- ========
-
- Go to the `Examples wiki <http://code.google.com/p/urllib3/wiki/Examples>`_
- for more nice syntax-highlighted examples.
-
- But, long story short::
-
- import urllib3
-
- API_URL = 'http://ajax.googleapis.com/ajax/services/search/web'
-
- http_pool = urllib3.connection_from_url(API_URL)
-
- fields = {'v': '1.0', 'q': 'urllib3'}
- r = http_pool.get_url(API_URL, fields)
-
- print r.status, r.data
-
-Keywords: urllib httplib threadsafe filepost http
-Platform: UNKNOWN
-Classifier: Environment :: Web Environment
-Classifier: Intended Audience :: Developers
-Classifier: License :: OSI Approved :: MIT License
-Classifier: Operating System :: OS Independent
-Classifier: Programming Language :: Python
-Classifier: Topic :: Internet :: WWW/HTTP
-Classifier: Topic :: Software Development :: Libraries
@@ -1,17 +0,0 @@
-README.txt
-setup.py
-test/benchmark.py
-test/dummy_server.py
-test/test_connectionpool.py
-test/test_withdummy.py
-urllib3/__init__.py
-urllib3/connectionpool.py
-urllib3/filepost.py
-urllib3.egg-info/PKG-INFO
-urllib3.egg-info/SOURCES.txt
-urllib3.egg-info/dependency_links.txt
-urllib3.egg-info/entry_points.txt
-urllib3.egg-info/not-zip-safe
-urllib3.egg-info/top_level.txt
-urllib3/contrib/__init__.py
-urllib3/contrib/ntlmpool.py
@@ -1 +0,0 @@
-
@@ -1,3 +0,0 @@
-
- # -*- Entry points: -*-
-
@@ -1 +0,0 @@
-
@@ -1 +0,0 @@
-urllib3

0 comments on commit edda198

Please sign in to comment.