diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 56fb24bc0..62fe2ab6c 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -8,7 +8,7 @@ repos: - id: check-added-large-files args: ["--maxkb=4000"] - repo: https://github.com/psf/black-pre-commit-mirror - rev: 25.11.0 + rev: 26.1.0 hooks: - id: black - repo: https://github.com/pycqa/isort @@ -33,7 +33,7 @@ repos: # additional_dependencies: [cpplint==1.6.1] # types_or: [c++] - repo: https://github.com/astral-sh/ruff-pre-commit - rev: v0.14.5 + rev: v0.14.14 hooks: - id: ruff args: [ "--fix", "--config", "ruff.toml" ] diff --git a/doc/sphinx_util.py b/doc/sphinx_util.py index cbdc3cb52..0608b96de 100644 --- a/doc/sphinx_util.py +++ b/doc/sphinx_util.py @@ -1,5 +1,6 @@ # -*- coding: utf-8 -*- """Helper utilty function for customization.""" + import os import subprocess import sys diff --git a/tracker/dmlc_tracker/kubernetes.py b/tracker/dmlc_tracker/kubernetes.py index 371d3d976..aeab92a5e 100644 --- a/tracker/dmlc_tracker/kubernetes.py +++ b/tracker/dmlc_tracker/kubernetes.py @@ -4,6 +4,7 @@ One need to make sure kubectl-able. """ + from __future__ import absolute_import import yaml diff --git a/tracker/dmlc_tracker/launcher.py b/tracker/dmlc_tracker/launcher.py index 36eb89344..dfe1aab58 100755 --- a/tracker/dmlc_tracker/launcher.py +++ b/tracker/dmlc_tracker/launcher.py @@ -1,6 +1,7 @@ #!/usr/bin/env python3 # pylint: disable=invalid-name """The container launcher script that launches DMLC with the right env variable.""" + from __future__ import absolute_import import glob @@ -55,7 +56,7 @@ def main(): if hadoop_home: library_path.append("%s/lib/native" % hdfs_home) library_path.append("%s/lib" % hdfs_home) - (classpath, _) = subprocess.Popen( + classpath, _ = subprocess.Popen( "%s/bin/hadoop classpath" % hadoop_home, stdout=subprocess.PIPE, shell=True, diff --git a/tracker/dmlc_tracker/mesos.py b/tracker/dmlc_tracker/mesos.py index 8714d410b..a089fc144 100644 --- a/tracker/dmlc_tracker/mesos.py +++ b/tracker/dmlc_tracker/mesos.py @@ -4,6 +4,7 @@ One need to make sure all slaves machines are ssh-able. """ + from __future__ import absolute_import import json diff --git a/tracker/dmlc_tracker/mpi.py b/tracker/dmlc_tracker/mpi.py index 1c4e5acc2..a5fbd8842 100644 --- a/tracker/dmlc_tracker/mpi.py +++ b/tracker/dmlc_tracker/mpi.py @@ -26,7 +26,7 @@ def get_mpi_env(envs): return cmd # decide MPI version. - (out, err) = subprocess.Popen( + out, err = subprocess.Popen( ["mpirun", "--version"], stdout=subprocess.PIPE, stderr=subprocess.PIPE ).communicate() if b"Open MPI" in out: diff --git a/tracker/dmlc_tracker/opts.py b/tracker/dmlc_tracker/opts.py index 907a11e6f..049e6b78c 100644 --- a/tracker/dmlc_tracker/opts.py +++ b/tracker/dmlc_tracker/opts.py @@ -1,5 +1,6 @@ # pylint: disable=invalid-name """Command line options of job submission script.""" + import argparse import os @@ -317,7 +318,7 @@ def get_opts(args=None): type=int, help=("Number of attempt local tracker can restart slave."), ) - (args, unknown) = parser.parse_known_args(args) + args, unknown = parser.parse_known_args(args) args.command += unknown if args.cluster is None: diff --git a/tracker/dmlc_tracker/ssh.py b/tracker/dmlc_tracker/ssh.py index 4ca72a895..37c7885d3 100644 --- a/tracker/dmlc_tracker/ssh.py +++ b/tracker/dmlc_tracker/ssh.py @@ -4,6 +4,7 @@ One need to make sure all slaves machines are ssh-able. """ + from __future__ import absolute_import import logging @@ -98,7 +99,7 @@ def run(prog): # launch jobs for i in range(nworker + nserver): pass_envs["DMLC_ROLE"] = "server" if i < nserver else "worker" - (node, port) = hosts[i % len(hosts)] + node, port = hosts[i % len(hosts)] pass_envs["DMLC_NODE_HOST"] = node prog = ( get_env(pass_envs) diff --git a/tracker/dmlc_tracker/yarn.py b/tracker/dmlc_tracker/yarn.py index 7547de275..a4d34f2ca 100755 --- a/tracker/dmlc_tracker/yarn.py +++ b/tracker/dmlc_tracker/yarn.py @@ -55,13 +55,13 @@ def yarn_submit(args, nworker, nserver, pass_env): ), "failed to build dmlc-yarn.jar, try it manually" # detech hadoop version - (out, _) = subprocess.Popen( + out, _ = subprocess.Popen( "%s version" % hadoop_binary, shell=True, stdout=subprocess.PIPE ).communicate() out = py_str(out).split("\n")[0].split() assert out[0] == "Hadoop", "cannot parse hadoop version string" hadoop_version = int(out[1].split(".")[0]) - (classpath, _) = subprocess.Popen( + classpath, _ = subprocess.Popen( "%s classpath" % hadoop_binary, shell=True, stdout=subprocess.PIPE ).communicate() classpath = py_str(classpath).strip()