Permalink
Browse files

Fix spelling mistakes

git ls-files | misspellings -f -
Source: https://github.com/lyda/misspell-check

Change-Id: I4132e6a276e44e2a8985238358533d315ee8d9c4
  • Loading branch information...
1 parent af3bd46 commit 45f0502b525de72ff98b118cfbbb9627c7caf165 @jogo jogo committed Feb 12, 2013
View
@@ -183,7 +183,7 @@ swift (1.6.0)
substantially affects the JSON output of the dispersion report, and any
tools written to consume this output will need to be updated.
- * Added Solaris (Illumos) compability
+ * Added Solaris (Illumos) compatibility
* Added -a option to swift-get-nodes to show all handoffs
View
@@ -106,7 +106,7 @@ class Auditor(object):
if resp.getheader('ETag').strip('"') != calc_hash:
self.object_checksum_mismatch += 1
consistent = False
- print ' MD5 doesnt match etag for "%s" on %s/%s' \
+ print ' MD5 does not match etag for "%s" on %s/%s' \
% (path, node['ip'], node['device'])
etags.append(resp.getheader('ETag'))
else:
@@ -354,7 +354,7 @@ Request URI Description
/recon/sockstat returns consumable info from /proc/net/sockstat|6
/recon/devices returns list of devices and devices dir i.e. /srv/node
/recon/async returns count of async pending
-/recon/replication returns object replication times (for backward compatability)
+/recon/replication returns object replication times (for backward compatibility)
/recon/replication/<type> returns replication info for given type (account, container, object)
/recon/auditor/<type> returns auditor stats on last reported scan for given type (account, container, object)
/recon/updater/<type> returns last updater sweep times for given type (container, object)
View
@@ -24,12 +24,12 @@ The supported headers are,
+----------------------------------------------+------------------------------+
|X-Container-Meta-Access-Control-Allow-Headers | Headers to be allowed in |
| | actual request by browser, |
-| | space seperated. |
+| | space separated. |
+----------------------------------------------+------------------------------+
|X-Container-Meta-Access-Control-Expose-Headers| Headers exposed to the user |
| | agent (e.g. browser) in the |
| | the actual request response. |
-| | Space seperated. |
+| | Space separated. |
+----------------------------------------------+------------------------------+
Before a browser issues an actual request it may issue a `preflight request`_.
@@ -70,7 +70,7 @@ header,::
-H 'X-Container-Meta-Access-Control-Allow-Origin: http://localhost' \
http://192.168.56.3:8080/v1/AUTH_test/cont1
-At this point the container is now accessable to CORS clients hosted on
+At this point the container is now accessible to CORS clients hosted on
http://localhost. Open the test CORS page in your browser.
#. Populate the Token field
@@ -153,7 +153,8 @@ def check_mount(root, drive):
"""
Verify that the path to the device is a mount point and mounted. This
allows us to fast fail on drives that have been unmounted because of
- issues, and also prevents us for accidently filling up the root partition.
+ issues, and also prevents us for accidentally filling up the root
+ partition.
:param root: base path where the devices are mounted
:param drive: drive name to be checked
View
@@ -793,7 +793,7 @@ def empty(self):
return (row[0] == 0)
def _commit_puts(self, item_list=None):
- """Handles commiting rows in .pending files."""
+ """Handles committing rows in .pending files."""
if self.db_file == ':memory:' or not os.path.exists(self.pending_file):
return
if item_list is None:
@@ -1320,7 +1320,7 @@ def _delete_db(self, conn, timestamp, force=False):
WHERE delete_timestamp < ? """, (timestamp, timestamp, timestamp))
def _commit_puts(self, item_list=None):
- """Handles commiting rows in .pending files."""
+ """Handles committing rows in .pending files."""
if self.db_file == ':memory:' or not os.path.exists(self.pending_file):
return
if item_list is None:
View
@@ -82,7 +82,7 @@ def wrapped(*a, **kw):
def watch_server_pids(server_pids, interval=1, **kwargs):
- """Monitor a collection of server pids yeilding back those pids that
+ """Monitor a collection of server pids yielding back those pids that
aren't responding to signals.
:param server_pids: a dict, lists of pids [int,...] keyed on
@@ -56,7 +56,7 @@
``X-Container-Meta-Web-Error: error.html`` will serve .../404error.html for
requests for paths not found.
-For psuedo paths that have no <index.name>, this middleware can serve HTML file
+For pseudo paths that have no <index.name>, this middleware can serve HTML file
listings if you set the ``X-Container-Meta-Web-Listings: true`` metadata item
on the container.
View
@@ -146,7 +146,7 @@ def __init__(self, noop=False):
self.func_name = 'posix_fallocate'
self.fallocate = noop_libc_function
return
- ## fallocate is prefered because we need the on-disk size to match
+ ## fallocate is preferred because we need the on-disk size to match
## the allocated size. Older versions of sqlite require that the
## two sizes match. However, fallocate is Linux only.
for func in ('fallocate', 'posix_fallocate'):
@@ -1066,7 +1066,7 @@ def compute_eta(start_time, current_value, final_value):
def iter_devices_partitions(devices_dir, item_type):
"""
- Iterate over partitions accross all devices.
+ Iterate over partitions across all devices.
:param devices_dir: Path to devices
:param item_type: One of 'accounts', 'containers', or 'objects'
View
@@ -295,7 +295,7 @@ def container_sync(self, path):
break
key = hash_path(info['account'], info['container'],
row['name'], raw_digest=True)
- # This node will only intially sync out one third of the
+ # This node will only initially sync out one third of the
# objects (if 3 replicas, 1/4 if 4, etc.). This section
# will attempt to sync previously skipped rows in case the
# other nodes didn't succeed.
@@ -313,7 +313,7 @@ def container_sync(self, path):
row = rows[0]
key = hash_path(info['account'], info['container'],
row['name'], raw_digest=True)
- # This node will only intially sync out one third of the
+ # This node will only initially sync out one third of the
# objects (if 3 replicas, 1/4 if 4, etc.). It'll come back
# around to the section above and attempt to sync
# previously skipped rows in case the other nodes didn't
@@ -408,7 +408,7 @@ def account_info(self, account, autocreate=False):
def container_info(self, account, container, account_autocreate=False):
"""
- Get container information and thusly verify container existance.
+ Get container information and thusly verify container existence.
This will also make a call to account_info to verify that the
account exists.
@@ -717,15 +717,15 @@ def write(self, data='', hdrs={}, parms={}, callback=None, cfg={}):
self.conn.put_start(self.path, hdrs=headers, parms=parms, cfg=cfg)
- transfered = 0
+ transferred = 0
buff = data.read(block_size)
try:
while len(buff) > 0:
self.conn.put_data(buff)
buff = data.read(block_size)
- transfered += len(buff)
+ transferred += len(buff)
if callable(callback):
- callback(transfered, self.size)
+ callback(transferred, self.size)
self.conn.put_end()
except socket.timeout, err:
@@ -611,7 +611,7 @@ def test_get_running_pids(self):
self.assertEquals(len(running_pids), 1)
self.assert_(1 in running_pids)
self.assert_(2 not in running_pids)
- # test persistant running pid files
+ # test persistent running pid files
self.assert_(os.path.exists(os.path.join(t, 'test-server1.pid')))
# test clean up stale pids
pid_two = self.join_swift_dir('test-server2.pid')
@@ -302,7 +302,7 @@ def test_LoggerFileObject(self):
lfo.tell()
def test_parse_options(self):
- # use mkstemp to get a file that is definately on disk
+ # use mkstemp to get a file that is definitely on disk
with NamedTemporaryFile() as f:
conf_file = f.name
conf, options = utils.parse_options(test_args=[conf_file])
@@ -154,7 +154,7 @@ def fake_audit_location_generator(*args, **kwargs):
def test_run_once(self):
# This runs runs_once with fakes twice, the first causing an interim
- # report, the second with no interm report.
+ # report, the second with no interim report.
time_calls = [0]
audit_location_generator_calls = [0]
@@ -373,7 +373,7 @@ def assertFileData(file_path, data):
whole_path_from = os.path.join(self.objects, '0', data_dir)
hashes_file = os.path.join(self.objects, '0',
object_replicator.HASH_FILE)
- # test that non existant file except caught
+ # test that non existent file except caught
self.assertEquals(object_replicator.invalidate_hash(whole_path_from),
None)
# test that hashes get cleared
@@ -1462,7 +1462,7 @@ def test_POST_meta_size(self):
limit = MAX_META_OVERALL_SIZE
controller = proxy_server.ObjectController(self.app, 'account',
'container', 'object')
- count = limit / 256 # enough to cause the limit to be reched
+ count = limit / 256 # enough to cause the limit to be reached
headers = dict(
(('X-Object-Meta-' + str(i), 'a' * 256)
for i in xrange(count + 1)))

0 comments on commit 45f0502

Please sign in to comment.