Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

run (parts of) pyupgrade to fix invalid escape sequences #2747

Merged
merged 1 commit into from
Jul 20, 2021
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
2 changes: 1 addition & 1 deletion nltk/app/chartparser_app.py
Original file line number Diff line number Diff line change
Expand Up @@ -2551,7 +2551,7 @@ def app():
for rule in grammar.productions():
print((" ", repr(rule) + ","))
print(")")
print(("tokens = %r" % tokens))
print("tokens = %r" % tokens)
print('Calling "ChartParserApp(grammar, tokens)"...')
ChartParserApp(grammar, tokens).mainloop()

Expand Down
14 changes: 6 additions & 8 deletions nltk/app/chunkparser_app.py
Original file line number Diff line number Diff line change
Expand Up @@ -1053,14 +1053,12 @@ def show_help(self, tab):
text = text.replace(
"<<TAGSET>>",
"\n".join(
(
"\t%s\t%s" % item
for item in sorted(
list(self.tagset.items()),
key=lambda t_w: re.match(r"\w+", t_w[0])
and (0, t_w[0])
or (1, t_w[0]),
)
"\t%s\t%s" % item
for item in sorted(
list(self.tagset.items()),
key=lambda t_w: re.match(r"\w+", t_w[0])
and (0, t_w[0])
or (1, t_w[0]),
)
),
)
Expand Down
14 changes: 5 additions & 9 deletions nltk/app/wordnet_app.py
Original file line number Diff line number Diff line change
Expand Up @@ -588,10 +588,8 @@ def _collect_all_synsets(word, pos, synset_relations=dict()):
part of speech.
"""
return "<ul>%s\n</ul>\n" % "".join(
(
_collect_one_synset(word, synset, synset_relations)
for synset in wn.synsets(word, pos)
)
_collect_one_synset(word, synset, synset_relations)
for synset in wn.synsets(word, pos)
)


Expand Down Expand Up @@ -647,11 +645,9 @@ def make_synset_html(db_name, disp_name, rels):
html = (
"<ul>"
+ "\n".join(
(
"<li>%s</li>" % make_synset_html(*rel_data)
for rel_data in get_relations_data(word, synset)
if rel_data[2] != []
)
"<li>%s</li>" % make_synset_html(*rel_data)
for rel_data in get_relations_data(word, synset)
if rel_data[2] != []
)
+ "</ul>"
)
Expand Down
2 changes: 1 addition & 1 deletion nltk/chunk/regexp.py
Original file line number Diff line number Diff line change
Expand Up @@ -1337,7 +1337,7 @@ def demo_eval(chunkparser, text):

print("/" + ("=" * 75) + "\\")
print("Scoring", chunkparser)
print(("-" * 77))
print("-" * 77)
print("Precision: %5.1f%%" %
(chunkscore.precision() * 100), " " * 4, end=" ")
print("Recall: %5.1f%%" % (chunkscore.recall() * 100), " " * 6, end=" ")
Expand Down
12 changes: 4 additions & 8 deletions nltk/classify/decisiontree.py
Original file line number Diff line number Diff line change
Expand Up @@ -268,10 +268,8 @@ def best_stump(feature_names, labeled_featuresets, verbose=False):
best_stump = stump
if verbose:
print(
(
"best stump for {:6d} toks uses {:20} err={:6.4f}".format(
len(labeled_featuresets), best_stump._fname, best_error
)
"best stump for {:6d} toks uses {:20} err={:6.4f}".format(
len(labeled_featuresets), best_stump._fname, best_error
)
)
return best_stump
Expand Down Expand Up @@ -322,10 +320,8 @@ def best_binary_stump(
else:
descr = "(default)"
print(
(
"best stump for {:6d} toks uses {:20} err={:6.4f}".format(
len(labeled_featuresets), descr, best_error
)
"best stump for {:6d} toks uses {:20} err={:6.4f}".format(
len(labeled_featuresets), descr, best_error
)
)
return best_stump
Expand Down
2 changes: 1 addition & 1 deletion nltk/classify/maxent.py
Original file line number Diff line number Diff line change
Expand Up @@ -1396,7 +1396,7 @@ def calculate_deltas(
deltas -= (ffreq_empirical - sum1) / -sum2

# We can stop once we converge.
n_error = numpy.sum(abs((ffreq_empirical - sum1))) / numpy.sum(abs(deltas))
n_error = numpy.sum(abs(ffreq_empirical - sum1)) / numpy.sum(abs(deltas))
if n_error < NEWTON_CONVERGE:
return deltas

Expand Down
6 changes: 2 additions & 4 deletions nltk/classify/naivebayes.py
Original file line number Diff line number Diff line change
Expand Up @@ -147,10 +147,8 @@ def labelprob(l):
cpdist[l1, fname].prob(fval) / cpdist[l0, fname].prob(fval)
)
print(
(
"%24s = %-14r %6s : %-6s = %s : 1.0"
% (fname, fval, ("%s" % l1)[:6], ("%s" % l0)[:6], ratio)
)
"%24s = %-14r %6s : %-6s = %s : 1.0"
% (fname, fval, ("%s" % l1)[:6], ("%s" % l0)[:6], ratio)
)

def most_informative_features(self, n=100):
Expand Down
2 changes: 1 addition & 1 deletion nltk/classify/senna.py
Original file line number Diff line number Diff line change
Expand Up @@ -132,7 +132,7 @@ def tag_sents(self, sentences):
_senna_cmd.extend(["-" + op for op in self.operations])

# Serialize the actual sentences to a temporary string
_input = "\n".join((" ".join(x) for x in sentences)) + "\n"
_input = "\n".join(" ".join(x) for x in sentences) + "\n"
if isinstance(_input, str) and encoding:
_input = _input.encode(encoding)

Expand Down
2 changes: 1 addition & 1 deletion nltk/classify/weka.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ def config_weka(classpath=None):
version = _check_weka_version(_weka_classpath)
if version:
print(
("[Found Weka: %s (version %s)]" % (_weka_classpath, version))
"[Found Weka: %s (version %s)]" % (_weka_classpath, version)
)
else:
print("[Found Weka: %s]" % _weka_classpath)
Expand Down
2 changes: 1 addition & 1 deletion nltk/data.py
Original file line number Diff line number Diff line change
Expand Up @@ -1294,7 +1294,7 @@ def tell(self):
bytes_read = (orig_filepos - len(self.bytebuffer)) - self._rewind_checkpoint
buf_size = sum(len(line) for line in self.linebuffer)
est_bytes = int(
(bytes_read * self._rewind_numchars / (self._rewind_numchars + buf_size))
bytes_read * self._rewind_numchars / (self._rewind_numchars + buf_size)
)

self.stream.seek(self._rewind_checkpoint)
Expand Down
6 changes: 3 additions & 3 deletions nltk/downloader.py
Original file line number Diff line number Diff line change
Expand Up @@ -1245,8 +1245,8 @@ def _show_config(self):
print()
print("Data Server:")
print(" - URL: <%s>" % self._ds.url)
print((" - %d Package Collections Available" % len(self._ds.collections())))
print((" - %d Individual Packages Available" % len(self._ds.packages())))
print(" - %d Package Collections Available" % len(self._ds.collections()))
print(" - %d Individual Packages Available" % len(self._ds.packages()))
print()
print("Local Machine:")
print(" - Data directory: %s" % self._ds.download_dir)
Expand All @@ -1268,7 +1268,7 @@ def _simple_interactive_config(self):
elif os.path.isdir(new_dl_dir):
self._ds.download_dir = new_dl_dir
else:
print(("Directory %r not found! Create it first." % new_dl_dir))
print("Directory %r not found! Create it first." % new_dl_dir)
elif user_input == "u":
new_url = input(" New URL> ").strip()
if new_url in ("", "x", "q", "X", "Q"):
Expand Down
2 changes: 1 addition & 1 deletion nltk/grammar.py
Original file line number Diff line number Diff line change
Expand Up @@ -754,7 +754,7 @@ def chomsky_normal_form(self, new_token_padding="@$@", flexible=False):
return self
if self.productions(empty=True):
raise ValueError(
("Grammar has Empty rules. " "Cannot deal with them at the moment")
"Grammar has Empty rules. " "Cannot deal with them at the moment"
)

# check for mixed rules
Expand Down
6 changes: 3 additions & 3 deletions nltk/metrics/agreement.py
Original file line number Diff line number Diff line change
Expand Up @@ -142,11 +142,11 @@ def agr(self, cA, cB, i, data=None):
# cfedermann: we don't know what combination of coder/item will come
# first in x; to avoid StopIteration problems due to assuming an order
# cA,cB, we allow either for k1 and then look up the missing as k2.
k1 = next((x for x in data if x["coder"] in (cA, cB) and x["item"] == i))
k1 = next(x for x in data if x["coder"] in (cA, cB) and x["item"] == i)
if k1["coder"] == cA:
k2 = next((x for x in data if x["coder"] == cB and x["item"] == i))
k2 = next(x for x in data if x["coder"] == cB and x["item"] == i)
else:
k2 = next((x for x in data if x["coder"] == cA and x["item"] == i))
k2 = next(x for x in data if x["coder"] == cA and x["item"] == i)

ret = 1.0 - float(self.distance(k1["labels"], k2["labels"]))
log.debug("Observed agreement between %s and %s on %s: %f", cA, cB, i, ret)
Expand Down
6 changes: 3 additions & 3 deletions nltk/metrics/distance.py
Original file line number Diff line number Diff line change
Expand Up @@ -250,11 +250,11 @@ def presence(label):

def fractional_presence(label):
return (
lambda x, y: abs(((1.0 / len(x)) - (1.0 / len(y))))
lambda x, y: abs((1.0 / len(x)) - (1.0 / len(y)))
* (label in x and label in y)
or 0.0 * (label not in x and label not in y)
or abs((1.0 / len(x))) * (label in x and label not in y)
or ((1.0 / len(y))) * (label not in x and label in y)
or abs(1.0 / len(x)) * (label in x and label not in y)
or (1.0 / len(y)) * (label not in x and label in y)
)


Expand Down
6 changes: 2 additions & 4 deletions nltk/misc/sort.py
Original file line number Diff line number Diff line change
Expand Up @@ -167,10 +167,8 @@ def demo():
count_quick = quick(a)

print(
(
("size=%5d: selection=%8d, bubble=%8d, " "merge=%6d, quick=%6d")
% (size, count_selection, count_bubble, count_merge, count_quick)
)
("size=%5d: selection=%8d, bubble=%8d, " "merge=%6d, quick=%6d")
% (size, count_selection, count_bubble, count_merge, count_quick)
)


Expand Down
12 changes: 5 additions & 7 deletions nltk/parse/nonprojectivedependencyparser.py
Original file line number Diff line number Diff line change
Expand Up @@ -150,13 +150,11 @@ def score(self, graph):
for head_node in graph.nodes.values():
for child_node in graph.nodes.values():
edges.append(
(
dict(
a=head_node["word"],
b=head_node["tag"],
c=child_node["word"],
d=child_node["tag"],
)
dict(
a=head_node["word"],
b=head_node["tag"],
c=child_node["word"],
d=child_node["tag"],
)
)

Expand Down
8 changes: 3 additions & 5 deletions nltk/parse/viterbi.py
Original file line number Diff line number Diff line change
Expand Up @@ -121,7 +121,7 @@ def parse(self, tokens):
# Initialize the constituents dictionary with the words from
# the text.
if self._trace:
print(("Inserting tokens into the most likely" + " constituents table..."))
print("Inserting tokens into the most likely" + " constituents table...")
for index in range(len(tokens)):
token = tokens[index]
constituents[index, index + 1, token] = token
Expand All @@ -133,10 +133,8 @@ def parse(self, tokens):
for length in range(1, len(tokens) + 1):
if self._trace:
print(
(
"Finding the most likely constituents"
+ " spanning %d text elements..." % length
)
"Finding the most likely constituents"
+ " spanning %d text elements..." % length
)
for start in range(len(tokens) - length + 1):
span = (start, start + length)
Expand Down
6 changes: 2 additions & 4 deletions nltk/probability.py
Original file line number Diff line number Diff line change
Expand Up @@ -2491,10 +2491,8 @@ def demo(numsamples=6, numoutcomes=500):

# Print the results in a formatted table.
print(
(
"%d samples (1-%d); %d outcomes were sampled for each FreqDist"
% (numsamples, numsamples, numoutcomes)
)
"%d samples (1-%d); %d outcomes were sampled for each FreqDist"
% (numsamples, numsamples, numoutcomes)
)
print("=" * 9 * (len(pdists) + 2))
FORMATSTR = " FreqDist " + "%8s " * (len(pdists) - 1) + "| Actual"
Expand Down
2 changes: 1 addition & 1 deletion nltk/sem/evaluate.py
Original file line number Diff line number Diff line change
Expand Up @@ -95,7 +95,7 @@ def set2rel(s):
if isinstance(elem, str):
new.add((elem,))
elif isinstance(elem, int):
new.add((str(elem)))
new.add(str(elem))
else:
new.add(elem)
return new
Expand Down
1 change: 0 additions & 1 deletion nltk/stem/rslp.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,3 @@

# Natural Language Toolkit: RSLP Stemmer
#
# Copyright (C) 2001-2021 NLTK Project
Expand Down
6 changes: 2 additions & 4 deletions nltk/stem/snowball.py
Original file line number Diff line number Diff line change
Expand Up @@ -5915,10 +5915,8 @@ def demo():

if language not in SnowballStemmer.languages:
print(
(
"\nOops, there is no stemmer for this language. "
+ "Please try again.\n"
)
"\nOops, there is no stemmer for this language. "
+ "Please try again.\n"
)
continue

Expand Down
2 changes: 1 addition & 1 deletion nltk/tag/brill_trainer.py
Original file line number Diff line number Diff line change
Expand Up @@ -279,7 +279,7 @@ def train(self, train_sents, max_rules=200, min_score=2, min_acc=None):
print("Finding initial useful rules...")
self._init_mappings(test_sents, train_sents)
if self._trace:
print((" Found {} useful rules.".format(len(self._rule_scores))))
print(" Found {} useful rules.".format(len(self._rule_scores)))

# Let the user know what we're up to.
if self._trace > 2:
Expand Down
4 changes: 2 additions & 2 deletions nltk/tag/senna.py
Original file line number Diff line number Diff line change
Expand Up @@ -105,8 +105,8 @@ def bio_to_chunks(self, tagged_sent, chunk_type):
for idx, word_pos in enumerate(tagged_sent):
word, pos = word_pos
if "-" + chunk_type in pos: # Append the word to the current_chunk.
current_chunk.append((word))
current_chunk_position.append((idx))
current_chunk.append(word)
current_chunk_position.append(idx)
else:
if current_chunk: # Flush the full chunk when out of an NP.
_chunk_str = " ".join(current_chunk)
Expand Down
2 changes: 1 addition & 1 deletion nltk/tag/stanford.py
Original file line number Diff line number Diff line change
Expand Up @@ -102,7 +102,7 @@ def tag_sents(self, sentences):

# Write the actual sentences to the temporary input file
_input_fh = os.fdopen(_input_fh, "wb")
_input = "\n".join((" ".join(x) for x in sentences))
_input = "\n".join(" ".join(x) for x in sentences)
if isinstance(_input, str) and encoding:
_input = _input.encode(encoding)
_input_fh.write(_input)
Expand Down
6 changes: 2 additions & 4 deletions nltk/tbl/feature.py
Original file line number Diff line number Diff line change
Expand Up @@ -220,10 +220,8 @@ def intersects(self, other):
"""

return bool(
(
self.__class__ is other.__class__
and set(self.positions) & set(other.positions)
)
self.__class__ is other.__class__
and set(self.positions) & set(other.positions)
)

# Rich comparisons for Features. With @functools.total_ordering (Python 2.7+),
Expand Down
1 change: 0 additions & 1 deletion nltk/test/childes_fixt.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,3 @@

def setup_module():
import pytest
import nltk.data
Expand Down
2 changes: 0 additions & 2 deletions nltk/test/classify_fixt.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@


# most of classify.doctest requires numpy
def setup_module():
import pytest
Expand Down
2 changes: 0 additions & 2 deletions nltk/test/discourse_fixt.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@


# FIXME: the entire discourse.doctest is skipped if Prover9/Mace4 is
# not installed, but there are pure-python parts that don't need Prover9.
def setup_module():
Expand Down
2 changes: 0 additions & 2 deletions nltk/test/gensim_fixt.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@


def setup_module():
import pytest
pytest.importorskip("gensim")
2 changes: 0 additions & 2 deletions nltk/test/gluesemantics_malt_fixt.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@


def setup_module():
import pytest
from nltk.parse.malt import MaltParser
Expand Down
2 changes: 0 additions & 2 deletions nltk/test/inference_fixt.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@


def setup_module():
import pytest
from nltk.inference.mace import Mace
Expand Down
2 changes: 0 additions & 2 deletions nltk/test/nonmonotonic_fixt.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@


def setup_module():
import pytest
from nltk.inference.mace import Mace
Expand Down
1 change: 0 additions & 1 deletion nltk/test/portuguese_en_fixt.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,3 @@

def setup_module():
import pytest

Expand Down
2 changes: 0 additions & 2 deletions nltk/test/probability_fixt.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@


# probability.doctest uses HMM which requires numpy;
# skip probability.doctest if numpy is not available

Expand Down