From 4f2ee8e0213a1307b5ddfc9689e7f18219e95f98 Mon Sep 17 00:00:00 2001 From: Reid 'arrdem' McKenzie Date: Mon, 30 Aug 2021 01:06:21 -0600 Subject: [PATCH] Done with flake8 --- .../src/python/anosql_migrations.py | 8 ++++---- .../test/python/test_migrations.py | 3 ++- projects/anosql/test/python/test_sqlite3.py | 8 +------- projects/calf/src/python/calf/cursedrepl.py | 8 ++++---- projects/calf/src/python/calf/parser.py | 4 ++-- .../src/python/datalog/shell/__main__.py | 13 +++++-------- .../datalog/src/python/datalog/evaluator.py | 2 +- projects/datalog/src/python/datalog/reader.py | 2 +- .../test/python/test_datalog_evaluator.py | 3 --- projects/flowmetal/scratch/astdump.py | 6 +++--- projects/flowmetal/scratch/astinterp.py | 8 +++++--- projects/jobq/benchmark.py | 2 +- projects/jobq/test/python/test_jobq.py | 10 +++++----- .../src/python/yamlschema/__init__.py | 10 +++++----- setup.cfg | 18 ++++++++++++++++-- tools/python/defs.bzl | 10 ---------- tools/python/test_licenses.py | 18 +++++++++--------- tools/sphinx/__main__.py | 4 ++-- 18 files changed, 66 insertions(+), 71 deletions(-) diff --git a/projects/anosql-migrations/src/python/anosql_migrations.py b/projects/anosql-migrations/src/python/anosql_migrations.py index a8fb8ca..425d130 100644 --- a/projects/anosql-migrations/src/python/anosql_migrations.py +++ b/projects/anosql-migrations/src/python/anosql_migrations.py @@ -131,16 +131,16 @@ def available_migrations(queries: Queries, conn) -> t.Iterable[MigrationDescript # query_fn: t.Callable + {.__name__, .__doc__, .sql} query_fn = getattr(queries, query_name) yield MigrationDescriptor( - name = query_name, - committed_at = None, - sha256sum = sha256(query_fn.sql.encode("utf-8")).hexdigest()) + name=query_name, + committed_at=None, + sha256sum=sha256(query_fn.sql.encode("utf-8")).hexdigest()) def execute_migration(queries: Queries, conn, migration: MigrationDescriptor): """Execute a given migration singularly.""" with conn: - # Mark the migration as in flight + # Mark the migration as in flight queries.anosql_migrations_create( conn, # Args diff --git a/projects/anosql-migrations/test/python/test_migrations.py b/projects/anosql-migrations/test/python/test_migrations.py index b4c6bae..afe83a8 100644 --- a/projects/anosql-migrations/test/python/test_migrations.py +++ b/projects/anosql-migrations/test/python/test_migrations.py @@ -13,6 +13,7 @@ _SQL = """\ CREATE TABLE kv (`id` INT, `key` TEXT, `value` TEXT); """ + def table_exists(conn, table_name): return list(conn.execute(f"""\ SELECT ( @@ -35,7 +36,7 @@ def conn() -> sqlite3.Connection: def test_connect(conn: sqlite3.Connection): """Assert that the connection works and we can execute against it.""" - assert list(conn.execute("SELECT 1;")) == [(1,),] + assert list(conn.execute("SELECT 1;")) == [(1, ), ] @pytest.fixture diff --git a/projects/anosql/test/python/test_sqlite3.py b/projects/anosql/test/python/test_sqlite3.py index 677c33a..df2ddf1 100644 --- a/projects/anosql/test/python/test_sqlite3.py +++ b/projects/anosql/test/python/test_sqlite3.py @@ -66,13 +66,7 @@ def test_insert_returning(sqlite3_conn, queries): ) print(blogid, type(blogid)) cur = sqlite3_conn.cursor() - cur.execute("""\ - select title - from blogs - where blogid = ?; - """, - (blogid,), - ) + cur.execute("SELECT `title` FROM `blogs` WHERE `blogid` = ?;", (blogid,)) actual = cur.fetchone() cur.close() expected = ("My first blog",) diff --git a/projects/calf/src/python/calf/cursedrepl.py b/projects/calf/src/python/calf/cursedrepl.py index 6de6683..dd20556 100644 --- a/projects/calf/src/python/calf/cursedrepl.py +++ b/projects/calf/src/python/calf/cursedrepl.py @@ -46,8 +46,8 @@ def curse_repl(handle_buffer): for ex, buff, vals, err in reversed(examples): putstr(f"Example {ex}:", attr=curses.A_BOLD) - for l in buff.split("\n"): - putstr(f" | {l}") + for line in buff.split("\n"): + putstr(f" | {line}") putstr("") @@ -55,8 +55,8 @@ def curse_repl(handle_buffer): err = str(err) err = err.split("\n") putstr(" Error:") - for l in err: - putstr(f" {l}", attr=curses.COLOR_YELLOW) + for line in err: + putstr(f" {line}", attr=curses.COLOR_YELLOW) elif vals: putstr(" Values:") diff --git a/projects/calf/src/python/calf/parser.py b/projects/calf/src/python/calf/parser.py index 8d9d469..cad30f8 100644 --- a/projects/calf/src/python/calf/parser.py +++ b/projects/calf/src/python/calf/parser.py @@ -24,9 +24,9 @@ def mk_sqlist(contents, open=None, close=None): ) -def pairwise(l: list) -> iter: +def pairwise(elems: list) -> iter: "s -> (s0,s1), (s2,s3), (s4, s5), ..." - return zip(l[::2], l[1::2]) + return zip(elems[::2], elems[1::2]) def mk_dict(contents, open=None, close=None): diff --git a/projects/datalog-shell/src/python/datalog/shell/__main__.py b/projects/datalog-shell/src/python/datalog/shell/__main__.py index 49cce32..b02d663 100755 --- a/projects/datalog-shell/src/python/datalog/shell/__main__.py +++ b/projects/datalog-shell/src/python/datalog/shell/__main__.py @@ -1,6 +1,4 @@ -#!/usr/bin/env python3 - -__doc__ = f""" +""" Datalog (py) ============ @@ -127,8 +125,7 @@ def main(args): db = db.merge(read_dataset(f.read())) print(f"Loaded {db_file} ...") except Exception as e: - print("Internal error - {e}") - print(f"Unable to load db {db_file}, skipping") + print(f"Internal error - {e}\nUnable to load db {db_file}, skipping") while True: try: @@ -155,8 +152,8 @@ def main(args): else: try: op, val = read_command(line) - except Exception as e: - print(f"Got an unknown command or syntax error, can't tell which") + except Exception: + print("Got an unknown command or a syntax error, can't tell which") continue # Definition merges on the DB @@ -215,7 +212,7 @@ def main(args): qdb = db.merge(db_cls([], [val])) val = val.pattern - with yaspin(SPINNER) as spinner: + with yaspin(SPINNER): with Timing() as t: try: results = list(select(qdb, val)) diff --git a/projects/datalog/src/python/datalog/evaluator.py b/projects/datalog/src/python/datalog/evaluator.py index 1f2de1c..05c8e62 100644 --- a/projects/datalog/src/python/datalog/evaluator.py +++ b/projects/datalog/src/python/datalog/evaluator.py @@ -19,7 +19,7 @@ def match(tuple, expr, bindings=None): # This may not work out long term. if isinstance(a, LVar) and isinstance(b, LVar): continue - elif isinstance(a, LVar) and not a in bindings and isinstance(b, Constant): + elif isinstance(a, LVar) and a not in bindings and isinstance(b, Constant): bindings[a] = b elif isinstance(a, LVar) and a in bindings and bindings[a] == b: continue diff --git a/projects/datalog/src/python/datalog/reader.py b/projects/datalog/src/python/datalog/reader.py index 2096823..1bc120c 100644 --- a/projects/datalog/src/python/datalog/reader.py +++ b/projects/datalog/src/python/datalog/reader.py @@ -4,7 +4,7 @@ A datalog reader. from collections import defaultdict -from datalog.parser import FAILURE, Grammar +from datalog.parser import FAILURE, Grammar, ParseError from datalog.types import Constant, Dataset, LVar, Rule diff --git a/projects/datalog/test/python/test_datalog_evaluator.py b/projects/datalog/test/python/test_datalog_evaluator.py index 3fa9850..2b328a3 100644 --- a/projects/datalog/test/python/test_datalog_evaluator.py +++ b/projects/datalog/test/python/test_datalog_evaluator.py @@ -73,9 +73,6 @@ def test_lvar_unification(db_cls): def test_rule_join(db_cls): """Test a basic join query - the parent -> grandparent relation.""" - child = Constant("child") - gc = Constant("grandchild") - d = read( """ child(a, b). diff --git a/projects/flowmetal/scratch/astdump.py b/projects/flowmetal/scratch/astdump.py index 28c3e7d..4fba249 100644 --- a/projects/flowmetal/scratch/astdump.py +++ b/projects/flowmetal/scratch/astdump.py @@ -40,7 +40,7 @@ class TreeDumper(ast.NodeVisitor): self.visit(node) def visit(self, node): - nodetype = type(node) + # nodetype = type(node) nodename = node.__class__.__name__ indent = " " * len(self._stack) * 2 print(indent + nodename) @@ -59,14 +59,14 @@ class YAMLTreeDumper(ast.NodeVisitor): def node2yml(self, node): try: - nodetype = type(node) + # nodetype = type(node) nodename = node.__class__.__name__ return { "op": nodename, "props": {n: node.__dict__[n] for n in propnames(node)}, "children": [], } - except: + except Exception: print(repr(node), propnames(node), dir(node)) def visit(self, node): diff --git a/projects/flowmetal/scratch/astinterp.py b/projects/flowmetal/scratch/astinterp.py index f79616c..902bddd 100644 --- a/projects/flowmetal/scratch/astinterp.py +++ b/projects/flowmetal/scratch/astinterp.py @@ -1,3 +1,5 @@ +# flake8: noqa: all + # Python AST interpreter written in Python # # This module is part of the Pycopy https://github.com/pfalcon/pycopy @@ -105,6 +107,7 @@ def arg_name(arg): else: return arg.arg + def kwarg_defaults(args): if sys.version_info < (3, 0, 0): return args.defaults @@ -117,7 +120,6 @@ class TargetNonlocalFlow(Exception): a target application.""" - class TargetBreak(TargetNonlocalFlow): pass @@ -248,7 +250,7 @@ class ModuleInterpreter(StrictNodeVisitor): self.push_ns(ClassNS(node)) try: self.stmt_list_visit(node.body) - except: + except Exception: self.pop_ns() raise ns = self.ns @@ -563,7 +565,7 @@ class ModuleInterpreter(StrictNodeVisitor): it = iter(val) try: for elt_idx, t in enumerate(target.elts): - if getattr(ast, "Starred", None ) and isinstance(t, ast.Starred): + if getattr(ast, "Starred", None) and isinstance(t, ast.Starred): t = t.value all_elts = list(it) break_i = len(all_elts) - (len(target.elts) - elt_idx - 1) diff --git a/projects/jobq/benchmark.py b/projects/jobq/benchmark.py index 3646715..b07ab9a 100644 --- a/projects/jobq/benchmark.py +++ b/projects/jobq/benchmark.py @@ -159,7 +159,7 @@ if __name__ == "__main__": test_poll(q, reps) test_append(q, reps) - print(f"Testing with :memory:") + print("Testing with :memory:") q = JobQueue(":memory:") test_insert(q, reps) test_poll(q, reps) diff --git a/projects/jobq/test/python/test_jobq.py b/projects/jobq/test/python/test_jobq.py index 5779ff1..e13c0ba 100644 --- a/projects/jobq/test/python/test_jobq.py +++ b/projects/jobq/test/python/test_jobq.py @@ -47,8 +47,8 @@ def test_poll(db): j1 = db.create("payload 1") j2 = db.create("payload 2") assert j1.modified == j2.modified, "Two within the second to force the `rowid` ASC" - sleep(1) # And a side-effect for the third one - j3 = db.create("payload 3") + sleep(1) # And a side-effect for the third one + db.create("payload 3") j = db.poll("true", ["assigned"]) @@ -60,7 +60,7 @@ def test_poll(db): def test_poll_not_found(db): """Test that poll can return nothing.""" - j1 = db.create("payload 1") + db.create("payload 1") j = db.poll("false", ["assigned"]) assert j is None @@ -69,7 +69,7 @@ def test_append(db, payload): """Test that appending an event to the log does append and preserves invariants.""" j = db.create(payload) - sleep(1) ## side-effect so that sqlite3 gets a different commit timestamp + sleep(1) # side-effect so that sqlite3 gets a different commit timestamp j_prime = db.append_event(j.id, "some user-defined event") assert isinstance(j_prime, Job) @@ -85,7 +85,7 @@ def test_cas_ok(db): """Test that we can CAS a job from one state to the 'next'.""" j = db.create("job2", ["state", 2]) - sleep(1) # side-effect so that sqlite3 gets a different commit timestamp + sleep(1) # side-effect so that sqlite3 gets a different commit timestamp j_prime = db.cas_state(j.id, ["state", 2], ["state", 3]) assert isinstance(j_prime, Job), "\n".join(db._db.iterdump()) diff --git a/projects/yamlschema/src/python/yamlschema/__init__.py b/projects/yamlschema/src/python/yamlschema/__init__.py index 3a192dd..70aa196 100644 --- a/projects/yamlschema/src/python/yamlschema/__init__.py +++ b/projects/yamlschema/src/python/yamlschema/__init__.py @@ -154,13 +154,13 @@ class YamlLinter(object): if maxl := schema.get("maxLength"): if len(node.value) > maxl: yield LintRecord( - LintLevel.MISSMATCH, node, schema, f"Expected a shorter string" + LintLevel.MISSMATCH, node, schema, "Expected a shorter string" ) if minl := schema.get("minLength"): if len(node.value) < minl: yield LintRecord( - LintLevel.MISSMATCH, node, schema, f"Expected a longer string" + LintLevel.MISSMATCH, node, schema, "Expected a longer string" ) if pat := schema.get("pattern"): @@ -169,7 +169,7 @@ class YamlLinter(object): LintLevel.MISSMATCH, node, schema, - f"Expected a string matching the pattern", + "Expected a string matching the pattern", ) def lint_integer(self, schema, node: Node) -> t.Iterable[LintRecord]: @@ -259,11 +259,11 @@ class YamlLinter(object): # Special schemas # These are schemas that accept everything. - if schema == True or schema == {}: + if schema is True or schema == {}: yield from [] # This is the schema that rejects everything. - elif schema == False: + elif schema is False: yield LintRecord( LintLevel.UNEXPECTED, node, schema, "Received an unexpected value" ) diff --git a/setup.cfg b/setup.cfg index 02ca290..c98f829 100644 --- a/setup.cfg +++ b/setup.cfg @@ -6,7 +6,12 @@ multi_line_output = 3 lines_after_imports = 2 default_section = THIRDPARTY known_localfolder = datalog -sections = FUTURE,STDLIB,LOCALFOLDER,THIRDPARTY +sections = [ + FUTURE, + STDLIB, + LOCALFOLDER, + THIRDPARTY, +] force_sort_within_sections = 1 force_alphabetical_sort_within_sections = 1 combine_as_imports = 1 @@ -18,7 +23,16 @@ index-servers = pypi shitlist = pip,pkg_resources,setuptools [flake8] -extend-ignore = E203,E501,F405,F403,E731,E306,E227 +extend-ignore = [ + E203, + E227, + E306, + E501, + E731, + E741, + F403, + F405, +] [pypi] repository = https://pypi.python.org/pypi diff --git a/tools/python/defs.bzl b/tools/python/defs.bzl index 1dc3f09..862685e 100644 --- a/tools/python/defs.bzl +++ b/tools/python/defs.bzl @@ -14,10 +14,6 @@ load("@rules_zapp//zapp:zapp.bzl", "zapp_binary", ) -load("//tools/flake8:flake8.bzl", - "flake8", -) - def py_requirement(*args, **kwargs): """A re-export of requirement()""" return _py_requirement(*args, **kwargs) @@ -211,12 +207,6 @@ def py_project(name=None, ], ) - # if lib_srcs: - # flake8( - # name = "flake8", - # deps = [lib_name], - # ) - if main: py_binary( name=name, diff --git a/tools/python/test_licenses.py b/tools/python/test_licenses.py index 524c99f..7aee264 100644 --- a/tools/python/test_licenses.py +++ b/tools/python/test_licenses.py @@ -58,7 +58,7 @@ LICENSES_BY_LOWERNAME = { # Mash in some cases. LICENSES_BY_LOWERNAME.update( - {l.split(" :: ")[-1].lower(): l for l in APPROVED_LICENSES} + {lic.split(" :: ")[-1].lower(): lic for lic in APPROVED_LICENSES} ) # As a workaround for packages which don"t have correct meadata on PyPi, hand-verified packages @@ -108,7 +108,7 @@ def licenses(package: Requirement): this problem. """ - l = [] + lics = [] version = next((v for op, v in package.specs if op == "=="), None) print(package.name, version) @@ -120,7 +120,7 @@ def licenses(package: Requirement): headers={"Accept": "application/json"} ).json() if ln := bash_license(blob.get("license")): - l.append(ln) + lics.append(ln) else: try: version = list(blob.get("releases", {}).keys())[-1] @@ -133,16 +133,16 @@ def licenses(package: Requirement): f"https://pypi.org/pypi/{package.name}/{version}/json", headers={"Accept": "application/json"} ).json() - l = [ + lics.extend([ c for c in blob.get("info", {}).get("classifiers", []) if c.startswith("License") - ] + ]) ln = blob.get("info", {}).get("license") - if ln and not l: - l.append(bash_license(ln)) + if ln and not lics: + lics.append(bash_license(ln)) - return l + return lics @pytest.mark.parametrize("package", PACKAGES) @@ -151,5 +151,5 @@ def test_approved_license(package): _licenses = licenses(package) assert package.name in APPROVED_PACKAGES or any( - l in APPROVED_LICENSES for l in _licenses + lic in APPROVED_LICENSES for lic in _licenses ), f"{package} was not approved and its license(s) were unknown {_licenses!r}" diff --git a/tools/sphinx/__main__.py b/tools/sphinx/__main__.py index d59f4ef..7f92055 100644 --- a/tools/sphinx/__main__.py +++ b/tools/sphinx/__main__.py @@ -105,7 +105,7 @@ def do_build( status = sys.stdout warning = sys.stderr - error = sys.stderr + # error = sys.stderr confdir = confdir or sourcedir confoverrides = {} # FIXME: support these @@ -176,7 +176,7 @@ def do_serve(host, port, sourcedir, outputdir): elif ( not path.startswith(outputdir) and path not in ignorelist - and not path in watchlist + and path not in watchlist ): # Watch any source file (file we open for reading) server.watch(path, build)