Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LINT: Fixup black string normalization #2929

Merged
merged 3 commits into from Aug 6, 2019
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
2 changes: 1 addition & 1 deletion distributed/cli/dask_scheduler.py
Expand Up @@ -215,7 +215,7 @@ def del_pid_file():
port=port,
dashboard_address=dashboard_address if dashboard else None,
service_kwargs={"dashboard": {"prefix": dashboard_prefix}},
**kwargs,
**kwargs
)
logger.info("Local Directory: %26s", local_directory)
logger.info("-" * 47)
Expand Down
2 changes: 1 addition & 1 deletion distributed/cli/dask_worker.py
Expand Up @@ -84,7 +84,7 @@
"--listen-address",
type=str,
default=None,
help="The address to which the worker binds. " "Example: tcp://0.0.0.0:9000",
help="The address to which the worker binds. Example: tcp://0.0.0.0:9000",
)
@click.option(
"--contact-address",
Expand Down
10 changes: 4 additions & 6 deletions distributed/client.py
Expand Up @@ -905,7 +905,7 @@ async def _start(self, timeout=no_default, **kwargs):
pass
except Exception:
logger.info(
"Tried to start cluster and received an error. " "Proceeding.",
"Tried to start cluster and received an error. Proceeding.",
exc_info=True,
)
address = self.cluster.scheduler_address
Expand Down Expand Up @@ -2383,9 +2383,7 @@ def _graph_to_futures(
dsk3 = {k: v for k, v in dsk2.items() if k is not v}
for future in extra_futures:
if future.client is not self:
msg = (
"Inputs contain futures that were created by " "another client."
)
msg = "Inputs contain futures that were created by another client."
raise ValueError(msg)

if restrictions:
Expand Down Expand Up @@ -3485,7 +3483,7 @@ def to_packages(d):
errs.append("%s\n%s" % (pkg, asciitable(["", "version"], rows)))

raise ValueError(
"Mismatched versions found\n" "\n" "%s" % ("\n\n".join(errs))
"Mismatched versions found\n\n%s" % ("\n\n".join(errs))
)

return result
Expand Down Expand Up @@ -3967,7 +3965,7 @@ async def _wait(fs, timeout=None, return_when=ALL_COMPLETED):
wait_for = Any
else:
raise NotImplementedError(
"Only return_when='ALL_COMPLETED' and 'FIRST_COMPLETED' are " "supported"
"Only return_when='ALL_COMPLETED' and 'FIRST_COMPLETED' are supported"
)

future = wait_for({f._state.wait() for f in fs})
Expand Down
2 changes: 1 addition & 1 deletion distributed/comm/addressing.py
Expand Up @@ -123,7 +123,7 @@ def get_address_host_port(addr, strict=False):
return backend.get_address_host_port(loc)
except NotImplementedError:
raise ValueError(
"don't know how to extract host and port " "for address %r" % (addr,)
"don't know how to extract host and port for address %r" % (addr,)
)


Expand Down
2 changes: 1 addition & 1 deletion distributed/comm/tests/test_ucx.py
Expand Up @@ -261,7 +261,7 @@ def test_ucx_localcluster(loop, processes):
threads_per_worker=1,
processes=processes,
loop=loop,
**kwargs,
**kwargs
) as cluster:
with Client(cluster) as client:
x = client.submit(inc, 1)
Expand Down
2 changes: 1 addition & 1 deletion distributed/diskutils.py
Expand Up @@ -237,7 +237,7 @@ def new_work_dir(self, **kwargs):
self._purge_leftovers()
except OSError:
logger.error(
"Failed to clean up lingering worker directories " "in path: %s ",
"Failed to clean up lingering worker directories in path: %s ",
exc_info=True,
)
return WorkDir(self, **kwargs)
2 changes: 1 addition & 1 deletion distributed/protocol/keras.py
Expand Up @@ -9,7 +9,7 @@ def serialize_keras_model(model):

if keras.__version__ < "1.2.0":
raise ImportError(
"Need Keras >= 1.2.0. " "Try pip install keras --upgrade --no-deps"
"Need Keras >= 1.2.0. Try pip install keras --upgrade --no-deps"
)

header = model._updated_config()
Expand Down
6 changes: 2 additions & 4 deletions distributed/scheduler.py
Expand Up @@ -2366,9 +2366,7 @@ def handle_long_running(self, key=None, worker=None, compute_duration=None):

ws = ts.processing_on
if ws is None:
logger.debug(
"Received long-running signal from duplicate task. " "Ignoring."
)
logger.debug("Received long-running signal from duplicate task. Ignoring.")
return

if compute_duration:
Expand Down Expand Up @@ -4730,7 +4728,7 @@ def check_worker_ttl(self):
for ws in self.workers.values():
if ws.last_seen < now - self.worker_ttl:
logger.warning(
"Worker failed to heartbeat within %s seconds. " "Closing: %s",
"Worker failed to heartbeat within %s seconds. Closing: %s",
self.worker_ttl,
ws,
)
Expand Down
2 changes: 1 addition & 1 deletion distributed/tests/test_worker.py
Expand Up @@ -464,7 +464,7 @@ def test_Executor(c, s):


@pytest.mark.skip(
reason="Other tests leak memory, so process-level checks" "trigger immediately"
reason="Other tests leak memory, so process-level checks trigger immediately"
)
@gen_cluster(
client=True,
Expand Down
2 changes: 1 addition & 1 deletion distributed/utils_perf.py
Expand Up @@ -42,7 +42,7 @@ def collect(self):
elapsed = max(collect_start - self.last_collect, MIN_RUNTIME)
if self.last_gc_duration / elapsed < self.max_in_gc_frac:
self.logger.debug(
"Calling gc.collect(). %0.3fs elapsed since " "previous call.", elapsed
"Calling gc.collect(). %0.3fs elapsed since previous call.", elapsed
)
gc.collect()
self.last_collect = collect_start
Expand Down
2 changes: 1 addition & 1 deletion distributed/worker.py
Expand Up @@ -3354,7 +3354,7 @@ async def run(server, comm, function, args=(), kwargs={}, is_coro=None, wait=Tru

except Exception as e:
logger.warning(
" Run Failed\n" "Function: %s\n" "args: %s\n" "kwargs: %s\n",
"Run Failed\nFunction: %s\nargs: %s\nkwargs: %s\n",
str(funcname(function))[:1000],
convert_args_to_str(args, max_len=1000),
convert_kwargs_to_str(kwargs, max_len=1000),
Expand Down