Skip to content

Upgrade Ruff to 0.11.0 #1890

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 5 commits into
base: master
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion backend/src/hatchling/builders/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -141,7 +141,7 @@ def include_spec(self) -> pathspec.GitIgnoreSpec | None:

# Matching only at the root requires a forward slash, back slashes do not work. As such,
# normalize to forward slashes for consistency.
all_include_patterns.extend(f"/{relative_path.replace(os.sep, '/')}/" for relative_path in self.packages)
all_include_patterns.extend(f'/{relative_path.replace(os.sep, "/")}/' for relative_path in self.packages)

if all_include_patterns:
return pathspec.GitIgnoreSpec.from_lines(all_include_patterns)
Expand Down
12 changes: 6 additions & 6 deletions backend/src/hatchling/builders/sdist.py
Original file line number Diff line number Diff line change
Expand Up @@ -230,15 +230,15 @@ def construct_setup_py_file(self, packages: list[str], extra_dependencies: tuple

authors_data = self.metadata.core.authors_data
if authors_data['name']:
contents += f" author={', '.join(authors_data['name'])!r},\n"
contents += f' author={", ".join(authors_data["name"])!r},\n'
if authors_data['email']:
contents += f" author_email={', '.join(authors_data['email'])!r},\n"
contents += f' author_email={", ".join(authors_data["email"])!r},\n'

maintainers_data = self.metadata.core.maintainers_data
if maintainers_data['name']:
contents += f" maintainer={', '.join(maintainers_data['name'])!r},\n"
contents += f' maintainer={", ".join(maintainers_data["name"])!r},\n'
if maintainers_data['email']:
contents += f" maintainer_email={', '.join(maintainers_data['email'])!r},\n"
contents += f' maintainer_email={", ".join(maintainers_data["email"])!r},\n'

if self.metadata.core.classifiers:
contents += ' classifiers=[\n'
Expand Down Expand Up @@ -313,9 +313,9 @@ def construct_setup_py_file(self, packages: list[str], extra_dependencies: tuple
for package in packages:
if package.startswith(f'src{os.sep}'):
src_layout = True
contents += f" {package.replace(os.sep, '.')[4:]!r},\n"
contents += f' {package.replace(os.sep, ".")[4:]!r},\n'
else:
contents += f" {package.replace(os.sep, '.')!r},\n"
contents += f' {package.replace(os.sep, ".")!r},\n'

contents += ' ],\n'

Expand Down
8 changes: 4 additions & 4 deletions backend/src/hatchling/builders/wheel.py
Original file line number Diff line number Diff line change
Expand Up @@ -483,7 +483,7 @@ def build_standard(self, directory: str, **build_data: Any) -> str:
records.write((f'{archive.metadata_directory}/RECORD', '', ''))
archive.write_metadata('RECORD', records.construct())

target = os.path.join(directory, f"{self.artifact_project_id}-{build_data['tag']}.whl")
target = os.path.join(directory, f'{self.artifact_project_id}-{build_data["tag"]}.whl')

replace_file(archive.path, target)
normalize_artifact_permissions(target)
Expand Down Expand Up @@ -572,7 +572,7 @@ def build_editable_detection(self, directory: str, **build_data: Any) -> str:
records.write((f'{archive.metadata_directory}/RECORD', '', ''))
archive.write_metadata('RECORD', records.construct())

target = os.path.join(directory, f"{self.artifact_project_id}-{build_data['tag']}.whl")
target = os.path.join(directory, f'{self.artifact_project_id}-{build_data["tag"]}.whl')

replace_file(archive.path, target)
normalize_artifact_permissions(target)
Expand All @@ -589,7 +589,7 @@ def build_editable_explicit(self, directory: str, **build_data: Any) -> str:
for relative_directory in self.config.dev_mode_dirs
)

record = archive.write_file(f"_{self.metadata.core.name.replace('-', '_')}.pth", '\n'.join(directories))
record = archive.write_file(f'_{self.metadata.core.name.replace("-", "_")}.pth', '\n'.join(directories))
records.write(record)

for included_file in self.recurse_forced_files(self.get_forced_inclusion_map(build_data)):
Expand All @@ -601,7 +601,7 @@ def build_editable_explicit(self, directory: str, **build_data: Any) -> str:
records.write((f'{archive.metadata_directory}/RECORD', '', ''))
archive.write_metadata('RECORD', records.construct())

target = os.path.join(directory, f"{self.artifact_project_id}-{build_data['tag']}.whl")
target = os.path.join(directory, f'{self.artifact_project_id}-{build_data["tag"]}.whl')

replace_file(archive.path, target)
normalize_artifact_permissions(target)
Expand Down
2 changes: 1 addition & 1 deletion backend/src/hatchling/cli/build/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ def build_impl(
builders[target_name] = builder_class

if unknown_targets:
app.abort(f"Unknown build targets: {', '.join(sorted(unknown_targets))}")
app.abort(f'Unknown build targets: {", ".join(sorted(unknown_targets))}')

# We guarantee that builds occur within the project directory
root = os.getcwd()
Expand Down
50 changes: 25 additions & 25 deletions backend/src/hatchling/metadata/spec.py
Original file line number Diff line number Diff line change
Expand Up @@ -214,15 +214,15 @@ def construct_metadata_file_1_2(metadata: ProjectMetadata, extra_dependencies: t

authors_data = metadata.core.authors_data
if authors_data['name']:
metadata_file += f"Author: {', '.join(authors_data['name'])}\n"
metadata_file += f'Author: {", ".join(authors_data["name"])}\n'
if authors_data['email']:
metadata_file += f"Author-email: {', '.join(authors_data['email'])}\n"
metadata_file += f'Author-email: {", ".join(authors_data["email"])}\n'

maintainers_data = metadata.core.maintainers_data
if maintainers_data['name']:
metadata_file += f"Maintainer: {', '.join(maintainers_data['name'])}\n"
metadata_file += f'Maintainer: {", ".join(maintainers_data["name"])}\n'
if maintainers_data['email']:
metadata_file += f"Maintainer-email: {', '.join(maintainers_data['email'])}\n"
metadata_file += f'Maintainer-email: {", ".join(maintainers_data["email"])}\n'

if metadata.core.license:
license_start = 'License: '
Expand All @@ -238,7 +238,7 @@ def construct_metadata_file_1_2(metadata: ProjectMetadata, extra_dependencies: t
metadata_file += f'License: {metadata.core.license_expression}\n'

if metadata.core.keywords:
metadata_file += f"Keywords: {','.join(metadata.core.keywords)}\n"
metadata_file += f'Keywords: {",".join(metadata.core.keywords)}\n'

if metadata.core.classifiers:
for classifier in metadata.core.classifiers:
Expand Down Expand Up @@ -275,15 +275,15 @@ def construct_metadata_file_2_1(metadata: ProjectMetadata, extra_dependencies: t

authors_data = metadata.core.authors_data
if authors_data['name']:
metadata_file += f"Author: {', '.join(authors_data['name'])}\n"
metadata_file += f'Author: {", ".join(authors_data["name"])}\n'
if authors_data['email']:
metadata_file += f"Author-email: {', '.join(authors_data['email'])}\n"
metadata_file += f'Author-email: {", ".join(authors_data["email"])}\n'

maintainers_data = metadata.core.maintainers_data
if maintainers_data['name']:
metadata_file += f"Maintainer: {', '.join(maintainers_data['name'])}\n"
metadata_file += f'Maintainer: {", ".join(maintainers_data["name"])}\n'
if maintainers_data['email']:
metadata_file += f"Maintainer-email: {', '.join(maintainers_data['email'])}\n"
metadata_file += f'Maintainer-email: {", ".join(maintainers_data["email"])}\n'

if metadata.core.license:
license_start = 'License: '
Expand All @@ -299,7 +299,7 @@ def construct_metadata_file_2_1(metadata: ProjectMetadata, extra_dependencies: t
metadata_file += f'License: {metadata.core.license_expression}\n'

if metadata.core.keywords:
metadata_file += f"Keywords: {','.join(metadata.core.keywords)}\n"
metadata_file += f'Keywords: {",".join(metadata.core.keywords)}\n'

if metadata.core.classifiers:
for classifier in metadata.core.classifiers:
Expand Down Expand Up @@ -361,15 +361,15 @@ def construct_metadata_file_2_2(metadata: ProjectMetadata, extra_dependencies: t

authors_data = metadata.core.authors_data
if authors_data['name']:
metadata_file += f"Author: {', '.join(authors_data['name'])}\n"
metadata_file += f'Author: {", ".join(authors_data["name"])}\n'
if authors_data['email']:
metadata_file += f"Author-email: {', '.join(authors_data['email'])}\n"
metadata_file += f'Author-email: {", ".join(authors_data["email"])}\n'

maintainers_data = metadata.core.maintainers_data
if maintainers_data['name']:
metadata_file += f"Maintainer: {', '.join(maintainers_data['name'])}\n"
metadata_file += f'Maintainer: {", ".join(maintainers_data["name"])}\n'
if maintainers_data['email']:
metadata_file += f"Maintainer-email: {', '.join(maintainers_data['email'])}\n"
metadata_file += f'Maintainer-email: {", ".join(maintainers_data["email"])}\n'

if metadata.core.license:
license_start = 'License: '
Expand All @@ -385,7 +385,7 @@ def construct_metadata_file_2_2(metadata: ProjectMetadata, extra_dependencies: t
metadata_file += f'License: {metadata.core.license_expression}\n'

if metadata.core.keywords:
metadata_file += f"Keywords: {','.join(metadata.core.keywords)}\n"
metadata_file += f'Keywords: {",".join(metadata.core.keywords)}\n'

if metadata.core.classifiers:
for classifier in metadata.core.classifiers:
Expand Down Expand Up @@ -447,15 +447,15 @@ def construct_metadata_file_2_3(metadata: ProjectMetadata, extra_dependencies: t

authors_data = metadata.core.authors_data
if authors_data['name']:
metadata_file += f"Author: {', '.join(authors_data['name'])}\n"
metadata_file += f'Author: {", ".join(authors_data["name"])}\n'
if authors_data['email']:
metadata_file += f"Author-email: {', '.join(authors_data['email'])}\n"
metadata_file += f'Author-email: {", ".join(authors_data["email"])}\n'

maintainers_data = metadata.core.maintainers_data
if maintainers_data['name']:
metadata_file += f"Maintainer: {', '.join(maintainers_data['name'])}\n"
metadata_file += f'Maintainer: {", ".join(maintainers_data["name"])}\n'
if maintainers_data['email']:
metadata_file += f"Maintainer-email: {', '.join(maintainers_data['email'])}\n"
metadata_file += f'Maintainer-email: {", ".join(maintainers_data["email"])}\n'

if metadata.core.license:
license_start = 'License: '
Expand All @@ -471,7 +471,7 @@ def construct_metadata_file_2_3(metadata: ProjectMetadata, extra_dependencies: t
metadata_file += f'License: {metadata.core.license_expression}\n'

if metadata.core.keywords:
metadata_file += f"Keywords: {','.join(metadata.core.keywords)}\n"
metadata_file += f'Keywords: {",".join(metadata.core.keywords)}\n'

if metadata.core.classifiers:
for classifier in metadata.core.classifiers:
Expand Down Expand Up @@ -533,15 +533,15 @@ def construct_metadata_file_2_4(metadata: ProjectMetadata, extra_dependencies: t

authors_data = metadata.core.authors_data
if authors_data['name']:
metadata_file += f"Author: {', '.join(authors_data['name'])}\n"
metadata_file += f'Author: {", ".join(authors_data["name"])}\n'
if authors_data['email']:
metadata_file += f"Author-email: {', '.join(authors_data['email'])}\n"
metadata_file += f'Author-email: {", ".join(authors_data["email"])}\n'

maintainers_data = metadata.core.maintainers_data
if maintainers_data['name']:
metadata_file += f"Maintainer: {', '.join(maintainers_data['name'])}\n"
metadata_file += f'Maintainer: {", ".join(maintainers_data["name"])}\n'
if maintainers_data['email']:
metadata_file += f"Maintainer-email: {', '.join(maintainers_data['email'])}\n"
metadata_file += f'Maintainer-email: {", ".join(maintainers_data["email"])}\n'

if metadata.core.license:
license_start = 'License: '
Expand All @@ -562,7 +562,7 @@ def construct_metadata_file_2_4(metadata: ProjectMetadata, extra_dependencies: t
metadata_file += f'License-File: {license_file}\n'

if metadata.core.keywords:
metadata_file += f"Keywords: {','.join(metadata.core.keywords)}\n"
metadata_file += f'Keywords: {",".join(metadata.core.keywords)}\n'

if metadata.core.classifiers:
for classifier in metadata.core.classifiers:
Expand Down
2 changes: 1 addition & 1 deletion backend/src/hatchling/version/source/code.py
Original file line number Diff line number Diff line change
Expand Up @@ -59,6 +59,6 @@ def get_version_data(self) -> dict:

return {'version': version}

def set_version(self, version: str, version_data: dict) -> None: # noqa: ARG002, PLR6301
def set_version(self, version: str, version_data: dict) -> None:
message = 'Cannot rewrite loaded code'
raise NotImplementedError(message)
2 changes: 1 addition & 1 deletion backend/src/hatchling/version/source/env.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,6 @@ def get_version_data(self) -> dict:

return {'version': os.environ[variable]}

def set_version(self, version: str, version_data: dict) -> None: # noqa: ARG002, PLR6301
def set_version(self, version: str, version_data: dict) -> None:
message = 'Cannot set environment variables'
raise NotImplementedError(message)
8 changes: 8 additions & 0 deletions docs/.hooks/render_ruff_defaults.py
Original file line number Diff line number Diff line change
Expand Up @@ -150,6 +150,9 @@ def run(self, lines): # noqa: PLR6301

UNSELECTED_RULES: tuple[str, ...] = (
'AIR001',
'AIR301',
'AIR302',
'AIR303',
'ANN001',
'ANN002',
'ANN003',
Expand Down Expand Up @@ -229,6 +232,7 @@ def run(self, lines): # noqa: PLR6301
'E305',
'E306',
'E501',
'E999',
'ERA001',
'FBT003',
'FIX001',
Expand Down Expand Up @@ -307,6 +311,8 @@ def run(self, lines): # noqa: PLR6301
'PTH205',
'PTH206',
'PTH207',
'PTH208',
'PTH210',
'Q000',
'Q001',
'Q002',
Expand All @@ -315,6 +321,7 @@ def run(self, lines): # noqa: PLR6301
'RET501',
'RET502',
'RUF011',
'RUF035',
'RUF200',
'S404',
'S410',
Expand All @@ -324,5 +331,6 @@ def run(self, lines): # noqa: PLR6301
'TD002',
'TD003',
'TRY200',
'UP027',
'W191',
)
1 change: 1 addition & 0 deletions docs/history/hatch.md
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,7 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),

***Added:***

- Upgrade Ruff to 0.11.0
- The `version` and `project metadata` commands now support projects that do not use Hatchling as the build backend
- The `version` command accepts a `--force` option, allowing for downgrades when an explicit version number is given.
- Build environments can now be configured, the default build environment is `hatch-build`
Expand Down
16 changes: 7 additions & 9 deletions ruff_defaults.toml
Original file line number Diff line number Diff line change
Expand Up @@ -500,7 +500,6 @@ select = [
"S317",
"S318",
"S319",
"S320",
"S321",
"S323",
"S324",
Expand Down Expand Up @@ -573,12 +572,12 @@ select = [
"T100",
"T201",
"T203",
"TCH001",
"TCH002",
"TCH003",
"TCH004",
"TCH005",
"TCH010",
"TC001",
"TC002",
"TC003",
"TC004",
"TC005",
"TC010",
"TD004",
"TD005",
"TD006",
Expand All @@ -590,9 +589,9 @@ select = [
"TRY003",
"TRY004",
"TRY201",
"TRY203",
"TRY300",
"TRY301",
"TRY302",
"TRY400",
"TRY401",
"UP001",
Expand Down Expand Up @@ -629,7 +628,6 @@ select = [
"UP035",
"UP036",
"UP037",
"UP038",
"UP039",
"UP040",
"UP041",
Expand Down
2 changes: 1 addition & 1 deletion scripts/generate_coverage_summary.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@

def main():
coverage_report = ROOT / 'coverage.xml'
root = etree.fromstring(coverage_report.read_text()) # nosec B320 # noqa: S320
root = etree.fromstring(coverage_report.read_text()) # nosec B320

raw_package_data = defaultdict(lambda: {'hits': 0, 'misses': 0})
for package in root.find('packages'):
Expand Down
3 changes: 1 addition & 2 deletions src/hatch/cli/build/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,8 +16,7 @@
'targets',
multiple=True,
help=(
'The target to build, overriding project defaults. '
'This may be selected multiple times e.g. `-t sdist -t wheel`'
'The target to build, overriding project defaults. This may be selected multiple times e.g. `-t sdist -t wheel`'
),
)
@click.option(
Expand Down
Loading
Loading