Compare commits

...

14 Commits
v1.8.1 ... main

Author SHA1 Message Date
Kevin Veen-Birkenbach
f4339a746a executet 'ruff format --check .'
Some checks are pending
Mark stable commit / test-unit (push) Waiting to run
Mark stable commit / test-integration (push) Waiting to run
Mark stable commit / test-env-virtual (push) Waiting to run
Mark stable commit / test-env-nix (push) Waiting to run
Mark stable commit / test-e2e (push) Waiting to run
Mark stable commit / test-virgin-user (push) Waiting to run
Mark stable commit / test-virgin-root (push) Waiting to run
Mark stable commit / lint-shell (push) Waiting to run
Mark stable commit / lint-python (push) Waiting to run
Mark stable commit / mark-stable (push) Blocked by required conditions
2025-12-18 14:04:44 +01:00
Kevin Veen-Birkenbach
763f02a9a4 Release version 1.8.6
Some checks failed
Mark stable commit / test-unit (push) Has been cancelled
Mark stable commit / test-integration (push) Has been cancelled
Mark stable commit / test-env-virtual (push) Has been cancelled
Mark stable commit / test-env-nix (push) Has been cancelled
Mark stable commit / test-e2e (push) Has been cancelled
Mark stable commit / test-virgin-user (push) Has been cancelled
Mark stable commit / test-virgin-root (push) Has been cancelled
Mark stable commit / lint-shell (push) Has been cancelled
Mark stable commit / lint-python (push) Has been cancelled
Mark stable commit / mark-stable (push) Has been cancelled
2025-12-17 23:50:31 +01:00
Kevin Veen-Birkenbach
2eec873a17 Solved Debian Bug
Some checks failed
Mark stable commit / test-unit (push) Has been cancelled
Mark stable commit / test-integration (push) Has been cancelled
Mark stable commit / test-env-virtual (push) Has been cancelled
Mark stable commit / test-env-nix (push) Has been cancelled
Mark stable commit / test-e2e (push) Has been cancelled
Mark stable commit / test-virgin-user (push) Has been cancelled
Mark stable commit / test-virgin-root (push) Has been cancelled
Mark stable commit / lint-shell (push) Has been cancelled
Mark stable commit / lint-python (push) Has been cancelled
Mark stable commit / mark-stable (push) Has been cancelled
https://chatgpt.com/share/69432655-a948-800f-8c0d-353921cdf644
2025-12-17 23:29:04 +01:00
Kevin Veen-Birkenbach
17ee947930 ci: pass NIX_CONFIG with GitHub token into all test containers
- Add NIX_CONFIG with GitHub access token to all CI test workflows
- Export NIX_CONFIG in Makefile for propagation to test scripts
- Forward NIX_CONFIG explicitly into all docker run invocations
- Prevent GitHub API rate limit errors during Nix-based tests

https://chatgpt.com/share/69432655-a948-800f-8c0d-353921cdf644
2025-12-17 23:29:04 +01:00
Kevin Veen-Birkenbach
b989bdd4eb Release version 1.8.5 2025-12-17 23:29:04 +01:00
Kevin Veen-Birkenbach
c4da8368d8 --- Release Error --- 2025-12-17 23:28:45 +01:00
Kevin Veen-Birkenbach
997c265cfb refactor(git): introduce GitRunError hierarchy, surface non-repo errors, and improve verification queries
Some checks failed
Mark stable commit / test-unit (push) Has been cancelled
Mark stable commit / test-integration (push) Has been cancelled
Mark stable commit / test-env-virtual (push) Has been cancelled
Mark stable commit / test-env-nix (push) Has been cancelled
Mark stable commit / test-e2e (push) Has been cancelled
Mark stable commit / test-virgin-user (push) Has been cancelled
Mark stable commit / test-virgin-root (push) Has been cancelled
Mark stable commit / lint-shell (push) Has been cancelled
Mark stable commit / lint-python (push) Has been cancelled
Mark stable commit / mark-stable (push) Has been cancelled
* Replace legacy GitError usage with a clearer exception hierarchy:

  * GitBaseError as the common root for all git-related failures
  * GitRunError for subprocess execution failures
  * GitQueryError for read-only query failures
  * GitCommandError for state-changing command failures
  * GitNotRepositoryError to explicitly signal “not a git repository” situations
* Update git runner to detect “not a git repository” stderr and raise GitNotRepositoryError with rich context (cwd, command, stderr)
* Refactor repository verification to use dedicated query helpers instead of ad-hoc subprocess calls:

  * get_remote_head_commit (ls-remote) for pull mode
  * get_head_commit for local mode
  * get_latest_signing_key (%GK) for signature verification
* Add strict vs best-effort behavior in verify_repository:

  * Best-effort collection for reporting (does not block when no verification config exists)
  * Strict retrieval and explicit error messages when verification is configured
  * Clear failure cases when commit/signing key cannot be determined
* Add new unit tests covering:

  * get_latest_signing_key output stripping and error wrapping
  * get_remote_head_commit parsing, empty output, and error wrapping
  * verify_repository success/failure scenarios and “do not swallow GitNotRepositoryError”
* Adjust imports and exception handling across actions/commands/queries to align with GitRunError-based handling while keeping GitNotRepositoryError uncaught for debugging clarity

https://chatgpt.com/share/6943173c-508c-800f-8879-af75d131c79b
2025-12-17 21:48:03 +01:00
Kevin Veen-Birkenbach
955028288f Release version 1.8.4
Some checks failed
Mark stable commit / test-unit (push) Has been cancelled
Mark stable commit / test-integration (push) Has been cancelled
Mark stable commit / test-env-virtual (push) Has been cancelled
Mark stable commit / test-env-nix (push) Has been cancelled
Mark stable commit / test-e2e (push) Has been cancelled
Mark stable commit / test-virgin-user (push) Has been cancelled
Mark stable commit / test-virgin-root (push) Has been cancelled
Mark stable commit / lint-shell (push) Has been cancelled
Mark stable commit / lint-python (push) Has been cancelled
Mark stable commit / mark-stable (push) Has been cancelled
2025-12-17 11:20:16 +01:00
Kevin Veen-Birkenbach
866572e252 ci(docker): fix repo mount path for pkgmgr as base layer of Infinito.Nexus
Some checks failed
Mark stable commit / test-unit (push) Has been cancelled
Mark stable commit / test-integration (push) Has been cancelled
Mark stable commit / test-env-virtual (push) Has been cancelled
Mark stable commit / test-env-nix (push) Has been cancelled
Mark stable commit / test-e2e (push) Has been cancelled
Mark stable commit / test-virgin-user (push) Has been cancelled
Mark stable commit / test-virgin-root (push) Has been cancelled
Mark stable commit / lint-shell (push) Has been cancelled
Mark stable commit / lint-python (push) Has been cancelled
Mark stable commit / mark-stable (push) Has been cancelled
Standardize Docker/CI/test environments to mount pkgmgr at /opt/src/pkgmgr.
This makes the layering explicit: pkgmgr is the lower-level foundation used by
Infinito.Nexus.

Infra-only change (Docker, CI, shell scripts). No runtime or Nix semantics changed.

https://chatgpt.com/share/69427fe7-e288-800f-90a4-c1c3c11a8484
2025-12-17 11:03:02 +01:00
Kevin Veen-Birkenbach
b0a733369e Optimized output for debugging
Some checks failed
Mark stable commit / test-unit (push) Has been cancelled
Mark stable commit / test-integration (push) Has been cancelled
Mark stable commit / test-env-virtual (push) Has been cancelled
Mark stable commit / test-env-nix (push) Has been cancelled
Mark stable commit / test-e2e (push) Has been cancelled
Mark stable commit / test-virgin-user (push) Has been cancelled
Mark stable commit / test-virgin-root (push) Has been cancelled
Mark stable commit / lint-shell (push) Has been cancelled
Mark stable commit / lint-python (push) Has been cancelled
Mark stable commit / mark-stable (push) Has been cancelled
2025-12-17 10:51:56 +01:00
Kevin Veen-Birkenbach
c5843ccd30 Release version 1.8.3
Some checks failed
Mark stable commit / test-unit (push) Has been cancelled
Mark stable commit / test-integration (push) Has been cancelled
Mark stable commit / test-env-virtual (push) Has been cancelled
Mark stable commit / test-env-nix (push) Has been cancelled
Mark stable commit / test-e2e (push) Has been cancelled
Mark stable commit / test-virgin-user (push) Has been cancelled
Mark stable commit / test-virgin-root (push) Has been cancelled
Mark stable commit / lint-shell (push) Has been cancelled
Mark stable commit / lint-python (push) Has been cancelled
Mark stable commit / mark-stable (push) Has been cancelled
2025-12-16 19:49:51 +01:00
Kevin Veen-Birkenbach
3cb7852cb4 feat(mirrors): support URL-only MIRRORS entries and keep git config clean
- Allow MIRRORS to contain plain URLs (one per line) in addition to legacy "NAME URL"
- Treat strings as single URLs to avoid iterable pitfalls
- Write PyPI URLs as metadata-only entries (never added to git config)
- Keep MIRRORS as the single source of truth for mirror setup
- Update integration test to assert URL-only MIRRORS output

https://chatgpt.com/share/6941a9aa-b8b4-800f-963d-2486b34856b1
2025-12-16 19:49:09 +01:00
Kevin Veen-Birkenbach
f995e3d368 Release version 1.8.2
Some checks failed
Mark stable commit / test-unit (push) Has been cancelled
Mark stable commit / test-integration (push) Has been cancelled
Mark stable commit / test-env-virtual (push) Has been cancelled
Mark stable commit / test-env-nix (push) Has been cancelled
Mark stable commit / test-e2e (push) Has been cancelled
Mark stable commit / test-virgin-user (push) Has been cancelled
Mark stable commit / test-virgin-root (push) Has been cancelled
Mark stable commit / lint-shell (push) Has been cancelled
Mark stable commit / lint-python (push) Has been cancelled
Mark stable commit / mark-stable (push) Has been cancelled
2025-12-16 19:22:41 +01:00
Kevin Veen-Birkenbach
ffa9d9660a gpt-5.2 ChatGPT: refactor tools code into cli.tools.vscode and add unit tests
Some checks failed
Mark stable commit / test-unit (push) Has been cancelled
Mark stable commit / test-integration (push) Has been cancelled
Mark stable commit / test-env-virtual (push) Has been cancelled
Mark stable commit / test-env-nix (push) Has been cancelled
Mark stable commit / test-e2e (push) Has been cancelled
Mark stable commit / test-virgin-user (push) Has been cancelled
Mark stable commit / test-virgin-root (push) Has been cancelled
Mark stable commit / lint-shell (push) Has been cancelled
Mark stable commit / lint-python (push) Has been cancelled
Mark stable commit / mark-stable (push) Has been cancelled
* Move VS Code workspace logic (incl. guards) from cli.commands.tools into cli.tools.vscode
* Extract shared repo path resolution into cli.tools.paths and reuse for explore/terminal
* Simplify cli.commands.tools to pure orchestration via open_vscode_workspace
* Update existing tools command unit test to assert delegation instead of patching removed internals
* Add new unit tests for cli.tools.paths and cli.tools.vscode (workspace creation, reuse, guard errors)

https://chatgpt.com/share/69419a6a-c9e4-800f-9538-b6652b2da6b3
2025-12-16 18:43:56 +01:00
222 changed files with 2222 additions and 1014 deletions

View File

@@ -11,7 +11,9 @@ jobs:
fail-fast: false fail-fast: false
matrix: matrix:
distro: [arch, debian, ubuntu, fedora, centos] distro: [arch, debian, ubuntu, fedora, centos]
env:
NIX_CONFIG: |
access-tokens = github.com=${{ secrets.GITHUB_TOKEN }}
steps: steps:
- name: Checkout repository - name: Checkout repository
uses: actions/checkout@v4 uses: actions/checkout@v4

View File

@@ -12,7 +12,9 @@ jobs:
fail-fast: false fail-fast: false
matrix: matrix:
distro: [arch, debian, ubuntu, fedora, centos] distro: [arch, debian, ubuntu, fedora, centos]
env:
NIX_CONFIG: |
access-tokens = github.com=${{ secrets.GITHUB_TOKEN }}
steps: steps:
- name: Checkout repository - name: Checkout repository
uses: actions/checkout@v4 uses: actions/checkout@v4

View File

@@ -11,7 +11,9 @@ jobs:
fail-fast: false fail-fast: false
matrix: matrix:
distro: [arch, debian, ubuntu, fedora, centos] distro: [arch, debian, ubuntu, fedora, centos]
env:
NIX_CONFIG: |
access-tokens = github.com=${{ secrets.GITHUB_TOKEN }}
steps: steps:
- name: Checkout repository - name: Checkout repository
uses: actions/checkout@v4 uses: actions/checkout@v4

View File

@@ -7,7 +7,9 @@ jobs:
test-integration: test-integration:
runs-on: ubuntu-latest runs-on: ubuntu-latest
timeout-minutes: 30 timeout-minutes: 30
env:
NIX_CONFIG: |
access-tokens = github.com=${{ secrets.GITHUB_TOKEN }}
steps: steps:
- name: Checkout repository - name: Checkout repository
uses: actions/checkout@v4 uses: actions/checkout@v4

View File

@@ -7,7 +7,9 @@ jobs:
test-unit: test-unit:
runs-on: ubuntu-latest runs-on: ubuntu-latest
timeout-minutes: 30 timeout-minutes: 30
env:
NIX_CONFIG: |
access-tokens = github.com=${{ secrets.GITHUB_TOKEN }}
steps: steps:
- name: Checkout repository - name: Checkout repository
uses: actions/checkout@v4 uses: actions/checkout@v4

View File

@@ -11,7 +11,9 @@ jobs:
fail-fast: false fail-fast: false
matrix: matrix:
distro: [arch, debian, ubuntu, fedora, centos] distro: [arch, debian, ubuntu, fedora, centos]
env:
NIX_CONFIG: |
access-tokens = github.com=${{ secrets.GITHUB_TOKEN }}
steps: steps:
- name: Checkout repository - name: Checkout repository
uses: actions/checkout@v4 uses: actions/checkout@v4
@@ -19,27 +21,26 @@ jobs:
- name: Show Docker version - name: Show Docker version
run: docker version run: docker version
# 🔹 BUILD virgin image if missing
- name: Build virgin container (${{ matrix.distro }}) - name: Build virgin container (${{ matrix.distro }})
run: | run: |
set -euo pipefail set -euo pipefail
PKGMGR_DISTRO="${{ matrix.distro }}" make build-missing-virgin PKGMGR_DISTRO="${{ matrix.distro }}" make build-missing-virgin
# 🔹 RUN test inside virgin image
- name: Virgin ${{ matrix.distro }} pkgmgr test (root) - name: Virgin ${{ matrix.distro }} pkgmgr test (root)
run: | run: |
set -euo pipefail set -euo pipefail
docker run --rm \ docker run --rm \
-v "$PWD":/src \ -v "$PWD":/opt/src/pkgmgr \
-v pkgmgr_repos:/root/Repositories \ -v pkgmgr_repos:/root/Repositories \
-v pkgmgr_pip_cache:/root/.cache/pip \ -v pkgmgr_pip_cache:/root/.cache/pip \
-w /src \ -e NIX_CONFIG="${NIX_CONFIG}" \
-w /opt/src/pkgmgr \
"pkgmgr-${{ matrix.distro }}-virgin" \ "pkgmgr-${{ matrix.distro }}-virgin" \
bash -lc ' bash -lc '
set -euo pipefail set -euo pipefail
git config --global --add safe.directory /src git config --global --add safe.directory /opt/src/pkgmgr
make install make install
make setup make setup
@@ -50,5 +51,5 @@ jobs:
pkgmgr version pkgmgr pkgmgr version pkgmgr
echo ">>> Running Nix-based: nix run .#pkgmgr -- version pkgmgr" echo ">>> Running Nix-based: nix run .#pkgmgr -- version pkgmgr"
nix run /src#pkgmgr -- version pkgmgr nix run /opt/src/pkgmgr#pkgmgr -- version pkgmgr
' '

View File

@@ -11,7 +11,9 @@ jobs:
fail-fast: false fail-fast: false
matrix: matrix:
distro: [arch, debian, ubuntu, fedora, centos] distro: [arch, debian, ubuntu, fedora, centos]
env:
NIX_CONFIG: |
access-tokens = github.com=${{ secrets.GITHUB_TOKEN }}
steps: steps:
- name: Checkout repository - name: Checkout repository
uses: actions/checkout@v4 uses: actions/checkout@v4
@@ -19,20 +21,19 @@ jobs:
- name: Show Docker version - name: Show Docker version
run: docker version run: docker version
# 🔹 BUILD virgin image if missing
- name: Build virgin container (${{ matrix.distro }}) - name: Build virgin container (${{ matrix.distro }})
run: | run: |
set -euo pipefail set -euo pipefail
PKGMGR_DISTRO="${{ matrix.distro }}" make build-missing-virgin PKGMGR_DISTRO="${{ matrix.distro }}" make build-missing-virgin
# 🔹 RUN test inside virgin image as non-root
- name: Virgin ${{ matrix.distro }} pkgmgr test (user) - name: Virgin ${{ matrix.distro }} pkgmgr test (user)
run: | run: |
set -euo pipefail set -euo pipefail
docker run --rm \ docker run --rm \
-v "$PWD":/src \ -v "$PWD":/opt/src/pkgmgr \
-w /src \ -e NIX_CONFIG="${NIX_CONFIG}" \
-w /opt/src/pkgmgr \
"pkgmgr-${{ matrix.distro }}-virgin" \ "pkgmgr-${{ matrix.distro }}-virgin" \
bash -lc ' bash -lc '
set -euo pipefail set -euo pipefail
@@ -42,7 +43,7 @@ jobs:
useradd -m dev useradd -m dev
echo "dev ALL=(ALL) NOPASSWD: ALL" > /etc/sudoers.d/dev echo "dev ALL=(ALL) NOPASSWD: ALL" > /etc/sudoers.d/dev
chmod 0440 /etc/sudoers.d/dev chmod 0440 /etc/sudoers.d/dev
chown -R dev:dev /src chown -R dev:dev /opt/src/pkgmgr
mkdir -p /nix/store /nix/var/nix /nix/var/log/nix /nix/var/nix/profiles mkdir -p /nix/store /nix/var/nix /nix/var/log/nix /nix/var/nix/profiles
chown -R dev:dev /nix chown -R dev:dev /nix
@@ -51,7 +52,7 @@ jobs:
sudo -H -u dev env HOME=/home/dev PKGMGR_DISABLE_NIX_FLAKE_INSTALLER=1 bash -lc " sudo -H -u dev env HOME=/home/dev PKGMGR_DISABLE_NIX_FLAKE_INSTALLER=1 bash -lc "
set -euo pipefail set -euo pipefail
cd /src cd /opt/src/pkgmgr
make setup-venv make setup-venv
. \"\$HOME/.venvs/pkgmgr/bin/activate\" . \"\$HOME/.venvs/pkgmgr/bin/activate\"
@@ -59,6 +60,6 @@ jobs:
pkgmgr version pkgmgr pkgmgr version pkgmgr
export NIX_REMOTE=local export NIX_REMOTE=local
nix run /src#pkgmgr -- version pkgmgr nix run /opt/src/pkgmgr#pkgmgr -- version pkgmgr
" "
' '

View File

@@ -1,3 +1,35 @@
## [1.8.6] - 2025-12-17
* Prevent Rate Limits during GitHub Nix Setups
## [1.8.5] - 2025-12-17
* * Clearer Git error handling, especially when a directory is not a Git repository.
* More reliable repository verification with improved commit and GPG signature checks.
* Better error messages and overall robustness when working with Git-based workflows.
## [1.9.0] - 2025-12-17
* Automated release.
## [1.8.4] - 2025-12-17
* * Made pkgmgrs base-layer role explicit by standardizing the Docker/CI mount path to *`/opt/src/pkgmgr`*.
## [1.8.3] - 2025-12-16
* MIRRORS now supports plain URL entries, ensuring metadata-only sources like PyPI are recorded without ever being added to the Git configuration.
## [1.8.2] - 2025-12-16
* * ***pkgmgr tools code*** is more robust and predictable: it now fails early with clear errors if VS Code is not installed or a repository is not yet identified.
## [1.8.1] - 2025-12-16 ## [1.8.1] - 2025-12-16
* * Improved stability and consistency of all Git operations (clone, pull, push, release, branch handling) with clearer error messages and predictable preview behavior. * * Improved stability and consistency of all Git operations (clone, pull, push, release, branch handling) with clearer error messages and predictable preview behavior.

View File

@@ -50,6 +50,6 @@ RUN set -euo pipefail; \
# Entry point # Entry point
COPY scripts/docker/entry.sh /usr/local/bin/docker-entry.sh COPY scripts/docker/entry.sh /usr/local/bin/docker-entry.sh
WORKDIR /src WORKDIR /opt/src/pkgmgr
ENTRYPOINT ["/usr/local/bin/docker-entry.sh"] ENTRYPOINT ["/usr/local/bin/docker-entry.sh"]
CMD ["pkgmgr", "--help"] CMD ["pkgmgr", "--help"]

View File

@@ -10,6 +10,10 @@ DISTROS ?= arch debian ubuntu fedora centos
PKGMGR_DISTRO ?= arch PKGMGR_DISTRO ?= arch
export PKGMGR_DISTRO export PKGMGR_DISTRO
# Nix Config Variable (To avoid rate limit)
NIX_CONFIG ?=
export NIX_CONFIG
# ------------------------------------------------------------ # ------------------------------------------------------------
# Base images # Base images
# (kept for documentation/reference; actual build logic is in scripts/build) # (kept for documentation/reference; actual build logic is in scripts/build)

View File

@@ -32,7 +32,7 @@
rec { rec {
pkgmgr = pyPkgs.buildPythonApplication { pkgmgr = pyPkgs.buildPythonApplication {
pname = "package-manager"; pname = "package-manager";
version = "1.8.1"; version = "1.8.6";
# Use the git repo as source # Use the git repo as source
src = ./.; src = ./.;

View File

@@ -1,7 +1,7 @@
# Maintainer: Kevin Veen-Birkenbach <info@veen.world> # Maintainer: Kevin Veen-Birkenbach <info@veen.world>
pkgname=package-manager pkgname=package-manager
pkgver=1.8.1 pkgver=1.8.6
pkgrel=1 pkgrel=1
pkgdesc="Local-flake wrapper for Kevin's package-manager (Nix-based)." pkgdesc="Local-flake wrapper for Kevin's package-manager (Nix-based)."
arch=('any') arch=('any')

View File

@@ -1,3 +1,41 @@
package-manager (1.8.6-1) unstable; urgency=medium
* Prevent Rate Limits during GitHub Nix Setups
-- Kevin Veen-Birkenbach <kevin@veen.world> Wed, 17 Dec 2025 23:50:31 +0100
package-manager (1.8.5-1) unstable; urgency=medium
* * Clearer Git error handling, especially when a directory is not a Git repository.
* More reliable repository verification with improved commit and GPG signature checks.
* Better error messages and overall robustness when working with Git-based workflows.
-- Kevin Veen-Birkenbach <kevin@veen.world> Wed, 17 Dec 2025 22:15:48 +0100
package-manager (1.9.0-1) unstable; urgency=medium
* Automated release.
-- Kevin Veen-Birkenbach <kevin@veen.world> Wed, 17 Dec 2025 22:10:31 +0100
package-manager (1.8.4-1) unstable; urgency=medium
* * Made pkgmgrs base-layer role explicit by standardizing the Docker/CI mount path to *`/opt/src/pkgmgr`*.
-- Kevin Veen-Birkenbach <kevin@veen.world> Wed, 17 Dec 2025 11:20:16 +0100
package-manager (1.8.3-1) unstable; urgency=medium
* MIRRORS now supports plain URL entries, ensuring metadata-only sources like PyPI are recorded without ever being added to the Git configuration.
-- Kevin Veen-Birkenbach <kevin@veen.world> Tue, 16 Dec 2025 19:49:51 +0100
package-manager (1.8.2-1) unstable; urgency=medium
* * ***pkgmgr tools code*** is more robust and predictable: it now fails early with clear errors if VS Code is not installed or a repository is not yet identified.
-- Kevin Veen-Birkenbach <kevin@veen.world> Tue, 16 Dec 2025 19:22:41 +0100
package-manager (1.8.1-1) unstable; urgency=medium package-manager (1.8.1-1) unstable; urgency=medium
* * Improved stability and consistency of all Git operations (clone, pull, push, release, branch handling) with clearer error messages and predictable preview behavior. * * Improved stability and consistency of all Git operations (clone, pull, push, release, branch handling) with clearer error messages and predictable preview behavior.

View File

@@ -1,5 +1,5 @@
Name: package-manager Name: package-manager
Version: 1.8.1 Version: 1.8.6
Release: 1%{?dist} Release: 1%{?dist}
Summary: Wrapper that runs Kevin's package-manager via Nix flake Summary: Wrapper that runs Kevin's package-manager via Nix flake
@@ -74,6 +74,26 @@ echo ">>> package-manager removed. Nix itself was not removed."
/usr/lib/package-manager/ /usr/lib/package-manager/
%changelog %changelog
* Wed Dec 17 2025 Kevin Veen-Birkenbach <kevin@veen.world> - 1.8.6-1
- Prevent Rate Limits during GitHub Nix Setups
* Wed Dec 17 2025 Kevin Veen-Birkenbach <kevin@veen.world> - 1.8.5-1
- * Clearer Git error handling, especially when a directory is not a Git repository.
* More reliable repository verification with improved commit and GPG signature checks.
* Better error messages and overall robustness when working with Git-based workflows.
* Wed Dec 17 2025 Kevin Veen-Birkenbach <kevin@veen.world> - 1.9.0-1
- Automated release.
* Wed Dec 17 2025 Kevin Veen-Birkenbach <kevin@veen.world> - 1.8.4-1
- * Made pkgmgrs base-layer role explicit by standardizing the Docker/CI mount path to *`/opt/src/pkgmgr`*.
* Tue Dec 16 2025 Kevin Veen-Birkenbach <kevin@veen.world> - 1.8.3-1
- MIRRORS now supports plain URL entries, ensuring metadata-only sources like PyPI are recorded without ever being added to the Git configuration.
* Tue Dec 16 2025 Kevin Veen-Birkenbach <kevin@veen.world> - 1.8.2-1
- * ***pkgmgr tools code*** is more robust and predictable: it now fails early with clear errors if VS Code is not installed or a repository is not yet identified.
* Tue Dec 16 2025 Kevin Veen-Birkenbach <kevin@veen.world> - 1.8.1-1 * Tue Dec 16 2025 Kevin Veen-Birkenbach <kevin@veen.world> - 1.8.1-1
- * Improved stability and consistency of all Git operations (clone, pull, push, release, branch handling) with clearer error messages and predictable preview behavior. - * Improved stability and consistency of all Git operations (clone, pull, push, release, branch handling) with clearer error messages and predictable preview behavior.
* Mirrors are now handled cleanly: only valid Git remotes are used for Git operations, while non-Git URLs (e.g. PyPI) are excluded, preventing broken or confusing repository configs. * Mirrors are now handled cleanly: only valid Git remotes are used for Git operations, while non-Git URLs (e.g. PyPI) are excluded, preventing broken or confusing repository configs.

View File

@@ -7,7 +7,7 @@ build-backend = "setuptools.build_meta"
[project] [project]
name = "kpmx" name = "kpmx"
version = "1.8.1" version = "1.8.6"
description = "Kevin's package-manager tool (pkgmgr)" description = "Kevin's package-manager tool (pkgmgr)"
readme = "README.md" readme = "README.md"
requires-python = ">=3.9" requires-python = ">=3.9"

View File

@@ -1,7 +1,7 @@
#!/usr/bin/env bash #!/usr/bin/env bash
set -euo pipefail set -euo pipefail
echo "[docker] Starting package-manager container" echo "[docker-pkgmgr] Starting package-manager container"
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
# Log distribution info # Log distribution info
@@ -9,19 +9,19 @@ echo "[docker] Starting package-manager container"
if [[ -f /etc/os-release ]]; then if [[ -f /etc/os-release ]]; then
# shellcheck disable=SC1091 # shellcheck disable=SC1091
. /etc/os-release . /etc/os-release
echo "[docker] Detected distro: ${ID:-unknown} (like: ${ID_LIKE:-})" echo "[docker-pkgmgr] Detected distro: ${ID:-unknown} (like: ${ID_LIKE:-})"
fi fi
# Always use /src (mounted from host) as working directory # Always use /opt/src/pkgmgr (mounted from host) as working directory
echo "[docker] Using /src as working directory" echo "[docker-pkgmgr] Using /opt/src/pkgmgr as working directory"
cd /src cd /opt/src/pkgmgr
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
# DEV mode: rebuild package-manager from the mounted /src tree # DEV mode: rebuild package-manager from the mounted /opt/src/pkgmgr tree
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
if [[ "${REINSTALL_PKGMGR:-0}" == "1" ]]; then if [[ "${REINSTALL_PKGMGR:-0}" == "1" ]]; then
echo "[docker] DEV mode enabled (REINSTALL_PKGMGR=1)" echo "[docker-pkgmgr] DEV mode enabled (REINSTALL_PKGMGR=1)"
echo "[docker] Rebuilding package-manager from /src via scripts/installation/package.sh..." echo "[docker-pkgmgr] Rebuilding package-manager from /opt/src/pkgmgr via scripts/installation/package.sh..."
bash scripts/installation/package.sh || exit 1 bash scripts/installation/package.sh || exit 1
fi fi
@@ -29,9 +29,9 @@ fi
# Hand off to pkgmgr or arbitrary command # Hand off to pkgmgr or arbitrary command
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
if [[ $# -eq 0 ]]; then if [[ $# -eq 0 ]]; then
echo "[docker] No arguments provided. Showing pkgmgr help..." echo "[docker-pkgmgr] No arguments provided. Showing pkgmgr help..."
exec pkgmgr --help exec pkgmgr --help
else else
echo "[docker] Executing command: $*" echo "[docker-pkgmgr] Executing command: $*"
exec "$@" exec "$@"
fi fi

View File

@@ -6,7 +6,7 @@ echo "[arch/package] Building Arch package (makepkg --nodeps) in an isolated bui
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)" SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
PROJECT_ROOT="$(cd "${SCRIPT_DIR}/../../.." && pwd)" PROJECT_ROOT="$(cd "${SCRIPT_DIR}/../../.." && pwd)"
# We must not build inside /src (mounted repo). Build in /tmp to avoid permission issues. # We must not build inside /opt/src/pkgmgr (mounted repo). Build in /tmp to avoid permission issues.
BUILD_ROOT="/tmp/package-manager-arch-build" BUILD_ROOT="/tmp/package-manager-arch-build"
PKG_SRC_DIR="${PROJECT_ROOT}/packaging/arch" PKG_SRC_DIR="${PROJECT_ROOT}/packaging/arch"
PKG_BUILD_DIR="${BUILD_ROOT}/packaging/arch" PKG_BUILD_DIR="${BUILD_ROOT}/packaging/arch"

View File

@@ -6,12 +6,13 @@ echo ">>> Running E2E tests: $PKGMGR_DISTRO"
echo "============================================================" echo "============================================================"
docker run --rm \ docker run --rm \
-v "$(pwd):/src" \ -v "$(pwd):/opt/src/pkgmgr" \
-v "pkgmgr_nix_store_${PKGMGR_DISTRO}:/nix" \ -v "pkgmgr_nix_store_${PKGMGR_DISTRO}:/nix" \
-v "pkgmgr_nix_cache_${PKGMGR_DISTRO}:/root/.cache/nix" \ -v "pkgmgr_nix_cache_${PKGMGR_DISTRO}:/root/.cache/nix" \
-e REINSTALL_PKGMGR=1 \ -e REINSTALL_PKGMGR=1 \
-e TEST_PATTERN="${TEST_PATTERN}" \ -e TEST_PATTERN="${TEST_PATTERN}" \
--workdir /src \ -e NIX_CONFIG="${NIX_CONFIG}" \
--workdir /opt/src/pkgmgr \
"pkgmgr-${PKGMGR_DISTRO}" \ "pkgmgr-${PKGMGR_DISTRO}" \
bash -lc ' bash -lc '
set -euo pipefail set -euo pipefail
@@ -40,14 +41,14 @@ docker run --rm \
} }
# Mark the mounted repository as safe to avoid Git ownership errors. # Mark the mounted repository as safe to avoid Git ownership errors.
# Newer Git (e.g. on Ubuntu) complains about the gitdir (/src/.git), # Newer Git (e.g. on Ubuntu) complains about the gitdir (/opt/src/pkgmgr/.git),
# older versions about the worktree (/src). Nix turns "." into the # older versions about the worktree (/opt/src/pkgmgr). Nix turns "." into the
# flake input "git+file:///src", which then uses Git under the hood. # flake input "git+file:///opt/src/pkgmgr", which then uses Git under the hood.
if command -v git >/dev/null 2>&1; then if command -v git >/dev/null 2>&1; then
# Worktree path # Worktree path
git config --global --add safe.directory /src || true git config --global --add safe.directory /opt/src/pkgmgr || true
# Gitdir path shown in the "dubious ownership" error # Gitdir path shown in the "dubious ownership" error
git config --global --add safe.directory /src/.git || true git config --global --add safe.directory /opt/src/pkgmgr/.git || true
# Ephemeral CI containers: allow all paths as a last resort # Ephemeral CI containers: allow all paths as a last resort
git config --global --add safe.directory "*" || true git config --global --add safe.directory "*" || true
fi fi
@@ -55,6 +56,6 @@ docker run --rm \
# Run the E2E tests inside the Nix development shell # Run the E2E tests inside the Nix development shell
nix develop .#default --no-write-lock-file -c \ nix develop .#default --no-write-lock-file -c \
python3 -m unittest discover \ python3 -m unittest discover \
-s /src/tests/e2e \ -s /opt/src/pkgmgr/tests/e2e \
-p "$TEST_PATTERN" -p "$TEST_PATTERN"
' '

View File

@@ -9,18 +9,19 @@ echo ">>> Image: ${IMAGE}"
echo "============================================================" echo "============================================================"
docker run --rm \ docker run --rm \
-v "$(pwd):/src" \ -v "$(pwd):/opt/src/pkgmgr" \
-v "pkgmgr_nix_store_${PKGMGR_DISTRO}:/nix" \ -v "pkgmgr_nix_store_${PKGMGR_DISTRO}:/nix" \
-v "pkgmgr_nix_cache_${PKGMGR_DISTRO}:/root/.cache/nix" \ -v "pkgmgr_nix_cache_${PKGMGR_DISTRO}:/root/.cache/nix" \
--workdir /src \ --workdir /opt/src/pkgmgr \
-e REINSTALL_PKGMGR=1 \ -e REINSTALL_PKGMGR=1 \
-e NIX_CONFIG="${NIX_CONFIG}" \
"${IMAGE}" \ "${IMAGE}" \
bash -lc ' bash -lc '
set -euo pipefail set -euo pipefail
if command -v git >/dev/null 2>&1; then if command -v git >/dev/null 2>&1; then
git config --global --add safe.directory /src || true git config --global --add safe.directory /opt/src/pkgmgr || true
git config --global --add safe.directory /src/.git || true git config --global --add safe.directory /opt/src/pkgmgr/.git || true
git config --global --add safe.directory "*" || true git config --global --add safe.directory "*" || true
fi fi
@@ -38,9 +39,9 @@ docker run --rm \
# ------------------------------------------------------------ # ------------------------------------------------------------
# Retry helper for GitHub API rate-limit (HTTP 403) # Retry helper for GitHub API rate-limit (HTTP 403)
# ------------------------------------------------------------ # ------------------------------------------------------------
if [[ -f /src/scripts/nix/lib/retry_403.sh ]]; then if [[ -f /opt/src/pkgmgr/scripts/nix/lib/retry_403.sh ]]; then
# shellcheck source=./scripts/nix/lib/retry_403.sh # shellcheck source=./scripts/nix/lib/retry_403.sh
source /src/scripts/nix/lib/retry_403.sh source /opt/src/pkgmgr/scripts/nix/lib/retry_403.sh
elif [[ -f ./scripts/nix/lib/retry_403.sh ]]; then elif [[ -f ./scripts/nix/lib/retry_403.sh ]]; then
# shellcheck source=./scripts/nix/lib/retry_403.sh # shellcheck source=./scripts/nix/lib/retry_403.sh
source ./scripts/nix/lib/retry_403.sh source ./scripts/nix/lib/retry_403.sh

View File

@@ -17,8 +17,9 @@ echo
# ------------------------------------------------------------ # ------------------------------------------------------------
if OUTPUT=$(docker run --rm \ if OUTPUT=$(docker run --rm \
-e REINSTALL_PKGMGR=1 \ -e REINSTALL_PKGMGR=1 \
-v "$(pwd):/src" \ -v "$(pwd):/opt/src/pkgmgr" \
-w /src \ -w /opt/src/pkgmgr \
-e NIX_CONFIG="${NIX_CONFIG}" \
"${IMAGE}" \ "${IMAGE}" \
bash -lc ' bash -lc '
set -euo pipefail set -euo pipefail

View File

@@ -6,19 +6,20 @@ echo ">>> Running INTEGRATION tests in ${PKGMGR_DISTRO} container"
echo "============================================================" echo "============================================================"
docker run --rm \ docker run --rm \
-v "$(pwd):/src" \ -v "$(pwd):/opt/src/pkgmgr" \
-v "pkgmgr_nix_store_${PKGMGR_DISTRO}:/nix" \ -v "pkgmgr_nix_store_${PKGMGR_DISTRO}:/nix" \
-v "pkgmgr_nix_cache_${PKGMGR_DISTRO}:/root/.cache/nix" \ -v "pkgmgr_nix_cache_${PKGMGR_DISTRO}:/root/.cache/nix" \
--workdir /src \ --workdir /opt/src/pkgmgr \
-e REINSTALL_PKGMGR=1 \ -e REINSTALL_PKGMGR=1 \
-e TEST_PATTERN="${TEST_PATTERN}" \ -e TEST_PATTERN="${TEST_PATTERN}" \
-e NIX_CONFIG="${NIX_CONFIG}" \
"pkgmgr-${PKGMGR_DISTRO}" \ "pkgmgr-${PKGMGR_DISTRO}" \
bash -lc ' bash -lc '
set -e; set -e;
git config --global --add safe.directory /src || true; git config --global --add safe.directory /opt/src/pkgmgr || true;
nix develop .#default --no-write-lock-file -c \ nix develop .#default --no-write-lock-file -c \
python3 -m unittest discover \ python3 -m unittest discover \
-s tests/integration \ -s tests/integration \
-t /src \ -t /opt/src/pkgmgr \
-p "$TEST_PATTERN"; -p "$TEST_PATTERN";
' '

View File

@@ -6,19 +6,20 @@ echo ">>> Running UNIT tests in ${PKGMGR_DISTRO} container"
echo "============================================================" echo "============================================================"
docker run --rm \ docker run --rm \
-v "$(pwd):/src" \ -v "$(pwd):/opt/src/pkgmgr" \
-v "pkgmgr_nix_cache_${PKGMGR_DISTRO}:/root/.cache/nix" \ -v "pkgmgr_nix_cache_${PKGMGR_DISTRO}:/root/.cache/nix" \
-v "pkgmgr_nix_store_${PKGMGR_DISTRO}:/nix" \ -v "pkgmgr_nix_store_${PKGMGR_DISTRO}:/nix" \
--workdir /src \ --workdir /opt/src/pkgmgr \
-e REINSTALL_PKGMGR=1 \ -e REINSTALL_PKGMGR=1 \
-e TEST_PATTERN="${TEST_PATTERN}" \ -e TEST_PATTERN="${TEST_PATTERN}" \
-e NIX_CONFIG="${NIX_CONFIG}" \
"pkgmgr-${PKGMGR_DISTRO}" \ "pkgmgr-${PKGMGR_DISTRO}" \
bash -lc ' bash -lc '
set -e; set -e;
git config --global --add safe.directory /src || true; git config --global --add safe.directory /opt/src/pkgmgr || true;
nix develop .#default --no-write-lock-file -c \ nix develop .#default --no-write-lock-file -c \
python3 -m unittest discover \ python3 -m unittest discover \
-s tests/unit \ -s tests/unit \
-t /src \ -t /opt/src/pkgmgr \
-p "$TEST_PATTERN"; -p "$TEST_PATTERN";
' '

View File

@@ -2,7 +2,7 @@ from __future__ import annotations
from typing import Optional from typing import Optional
from pkgmgr.core.git.errors import GitError from pkgmgr.core.git.errors import GitRunError
from pkgmgr.core.git.queries import get_current_branch from pkgmgr.core.git.queries import get_current_branch
from pkgmgr.core.git.commands import ( from pkgmgr.core.git.commands import (
GitDeleteRemoteBranchError, GitDeleteRemoteBranchError,
@@ -32,7 +32,7 @@ def close_branch(
if not name: if not name:
try: try:
name = get_current_branch(cwd=cwd) name = get_current_branch(cwd=cwd)
except GitError as exc: except GitRunError as exc:
raise RuntimeError(f"Failed to detect current branch: {exc}") from exc raise RuntimeError(f"Failed to detect current branch: {exc}") from exc
if not name: if not name:
@@ -48,14 +48,18 @@ def close_branch(
# Confirmation # Confirmation
if not force: if not force:
answer = input( answer = (
input(
f"Merge branch '{name}' into '{target_base}' and delete it afterwards? (y/N): " f"Merge branch '{name}' into '{target_base}' and delete it afterwards? (y/N): "
).strip().lower() )
.strip()
.lower()
)
if answer != "y": if answer != "y":
print("Aborted closing branch.") print("Aborted closing branch.")
return return
# Execute workflow (commands raise specific GitError subclasses) # Execute workflow (commands raise specific GitRunError subclasses)
fetch("origin", cwd=cwd) fetch("origin", cwd=cwd)
checkout(target_base, cwd=cwd) checkout(target_base, cwd=cwd)
pull("origin", target_base, cwd=cwd) pull("origin", target_base, cwd=cwd)

View File

@@ -2,7 +2,7 @@ from __future__ import annotations
from typing import Optional from typing import Optional
from pkgmgr.core.git.errors import GitError from pkgmgr.core.git.errors import GitRunError
from pkgmgr.core.git.queries import get_current_branch from pkgmgr.core.git.queries import get_current_branch
from pkgmgr.core.git.commands import ( from pkgmgr.core.git.commands import (
GitDeleteRemoteBranchError, GitDeleteRemoteBranchError,
@@ -26,7 +26,7 @@ def drop_branch(
if not name: if not name:
try: try:
name = get_current_branch(cwd=cwd) name = get_current_branch(cwd=cwd)
except GitError as exc: except GitRunError as exc:
raise RuntimeError(f"Failed to detect current branch: {exc}") from exc raise RuntimeError(f"Failed to detect current branch: {exc}") from exc
if not name: if not name:
@@ -41,9 +41,13 @@ def drop_branch(
# Confirmation # Confirmation
if not force: if not force:
answer = input( answer = (
input(
f"Delete branch '{name}' locally and on origin? This is destructive! (y/N): " f"Delete branch '{name}' locally and on origin? This is destructive! (y/N): "
).strip().lower() )
.strip()
.lower()
)
if answer != "y": if answer != "y":
print("Aborted dropping branch.") print("Aborted dropping branch.")
return return

View File

@@ -30,7 +30,7 @@ def open_branch(
resolved_base = resolve_base_branch(base_branch, fallback_base, cwd=cwd) resolved_base = resolve_base_branch(base_branch, fallback_base, cwd=cwd)
# Workflow (commands raise specific GitError subclasses) # Workflow (commands raise specific GitBaseError subclasses)
fetch("origin", cwd=cwd) fetch("origin", cwd=cwd)
checkout(resolved_base, cwd=cwd) checkout(resolved_base, cwd=cwd)
pull("origin", resolved_base, cwd=cwd) pull("origin", resolved_base, cwd=cwd)

View File

@@ -2,6 +2,7 @@ import yaml
import os import os
from pkgmgr.core.config.save import save_user_config from pkgmgr.core.config.save import save_user_config
def interactive_add(config, USER_CONFIG_PATH: str): def interactive_add(config, USER_CONFIG_PATH: str):
"""Interactively prompt the user to add a new repository entry to the user config.""" """Interactively prompt the user to add a new repository entry to the user config."""
print("Adding a new repository configuration entry.") print("Adding a new repository configuration entry.")
@@ -9,7 +10,9 @@ def interactive_add(config,USER_CONFIG_PATH:str):
new_entry["provider"] = input("Provider (e.g., github.com): ").strip() new_entry["provider"] = input("Provider (e.g., github.com): ").strip()
new_entry["account"] = input("Account (e.g., yourusername): ").strip() new_entry["account"] = input("Account (e.g., yourusername): ").strip()
new_entry["repository"] = input("Repository name (e.g., mytool): ").strip() new_entry["repository"] = input("Repository name (e.g., mytool): ").strip()
new_entry["command"] = input("Command (optional, leave blank to auto-detect): ").strip() new_entry["command"] = input(
"Command (optional, leave blank to auto-detect): "
).strip()
new_entry["description"] = input("Description (optional): ").strip() new_entry["description"] = input("Description (optional): ").strip()
new_entry["replacement"] = input("Replacement (optional): ").strip() new_entry["replacement"] = input("Replacement (optional): ").strip()
new_entry["alias"] = input("Alias (optional): ").strip() new_entry["alias"] = input("Alias (optional): ").strip()
@@ -25,7 +28,7 @@ def interactive_add(config,USER_CONFIG_PATH:str):
confirm = input("Add this entry to user config? (y/N): ").strip().lower() confirm = input("Add this entry to user config? (y/N): ").strip().lower()
if confirm == "y": if confirm == "y":
if os.path.exists(USER_CONFIG_PATH): if os.path.exists(USER_CONFIG_PATH):
with open(USER_CONFIG_PATH, 'r') as f: with open(USER_CONFIG_PATH, "r") as f:
user_config = yaml.safe_load(f) or {} user_config = yaml.safe_load(f) or {}
else: else:
user_config = {"repositories": []} user_config = {"repositories": []}

View File

@@ -107,11 +107,15 @@ def config_init(
# Already known? # Already known?
if key in default_keys: if key in default_keys:
skipped += 1 skipped += 1
print(f"[SKIP] (defaults) {provider}/{account}/{repo_name}") print(
f"[SKIP] (defaults) {provider}/{account}/{repo_name}"
)
continue continue
if key in existing_keys: if key in existing_keys:
skipped += 1 skipped += 1
print(f"[SKIP] (user-config) {provider}/{account}/{repo_name}") print(
f"[SKIP] (user-config) {provider}/{account}/{repo_name}"
)
continue continue
print(f"[ADD] {provider}/{account}/{repo_name}") print(f"[ADD] {provider}/{account}/{repo_name}")
@@ -121,7 +125,9 @@ def config_init(
if verified_commit: if verified_commit:
print(f"[INFO] Latest commit: {verified_commit}") print(f"[INFO] Latest commit: {verified_commit}")
else: else:
print("[WARN] Could not read commit (not a git repo or no commits).") print(
"[WARN] Could not read commit (not a git repo or no commits)."
)
entry: Dict[str, Any] = { entry: Dict[str, Any] = {
"provider": provider, "provider": provider,

View File

@@ -1,6 +1,7 @@
import yaml import yaml
from pkgmgr.core.config.load import load_config from pkgmgr.core.config.load import load_config
def show_config(selected_repos, user_config_path, full_config=False): def show_config(selected_repos, user_config_path, full_config=False):
"""Display configuration for one or more repositories, or the entire merged config.""" """Display configuration for one or more repositories, or the entire merged config."""
if full_config: if full_config:
@@ -8,7 +9,9 @@ def show_config(selected_repos, user_config_path, full_config=False):
print(yaml.dump(merged, default_flow_style=False)) print(yaml.dump(merged, default_flow_style=False))
else: else:
for repo in selected_repos: for repo in selected_repos:
identifier = f'{repo.get("provider")}/{repo.get("account")}/{repo.get("repository")}' identifier = (
f"{repo.get('provider')}/{repo.get('account')}/{repo.get('repository')}"
)
print(f"Repository: {identifier}") print(f"Repository: {identifier}")
for key, value in repo.items(): for key, value in repo.items():
print(f" {key}: {value}") print(f" {key}: {value}")

View File

@@ -66,10 +66,7 @@ def _ensure_repo_dir(
repo_dir = get_repo_dir(repositories_base_dir, repo) repo_dir = get_repo_dir(repositories_base_dir, repo)
if not os.path.exists(repo_dir): if not os.path.exists(repo_dir):
print( print(f"Repository directory '{repo_dir}' does not exist. Cloning it now...")
f"Repository directory '{repo_dir}' does not exist. "
"Cloning it now..."
)
clone_repos( clone_repos(
[repo], [repo],
repositories_base_dir, repositories_base_dir,
@@ -79,10 +76,7 @@ def _ensure_repo_dir(
clone_mode, clone_mode,
) )
if not os.path.exists(repo_dir): if not os.path.exists(repo_dir):
print( print(f"Cloning failed for repository {identifier}. Skipping installation.")
f"Cloning failed for repository {identifier}. "
"Skipping installation."
)
return None return None
return repo_dir return repo_dir
@@ -115,7 +109,9 @@ def _verify_repo(
if silent: if silent:
# Non-interactive mode: continue with a warning. # Non-interactive mode: continue with a warning.
print(f"[Warning] Continuing despite verification failure for {identifier} (--silent).") print(
f"[Warning] Continuing despite verification failure for {identifier} (--silent)."
)
else: else:
choice = input("Continue anyway? [y/N]: ").strip().lower() choice = input("Continue anyway? [y/N]: ").strip().lower()
if choice != "y": if choice != "y":
@@ -232,12 +228,16 @@ def install_repos(
code = exc.code if isinstance(exc.code, int) else str(exc.code) code = exc.code if isinstance(exc.code, int) else str(exc.code)
failures.append((identifier, f"installer failed (exit={code})")) failures.append((identifier, f"installer failed (exit={code})"))
if not quiet: if not quiet:
print(f"[Warning] install: repository {identifier} failed (exit={code}). Continuing...") print(
f"[Warning] install: repository {identifier} failed (exit={code}). Continuing..."
)
continue continue
except Exception as exc: except Exception as exc:
failures.append((identifier, f"unexpected error: {exc}")) failures.append((identifier, f"unexpected error: {exc}"))
if not quiet: if not quiet:
print(f"[Warning] install: repository {identifier} hit an unexpected error: {exc}. Continuing...") print(
f"[Warning] install: repository {identifier} hit an unexpected error: {exc}. Continuing..."
)
continue continue
if failures and emit_summary and not quiet: if failures and emit_summary and not quiet:

View File

@@ -14,6 +14,10 @@ from pkgmgr.actions.install.installers.python import PythonInstaller # noqa: F4
from pkgmgr.actions.install.installers.makefile import MakefileInstaller # noqa: F401 from pkgmgr.actions.install.installers.makefile import MakefileInstaller # noqa: F401
# OS-specific installers # OS-specific installers
from pkgmgr.actions.install.installers.os_packages.arch_pkgbuild import ArchPkgbuildInstaller # noqa: F401 from pkgmgr.actions.install.installers.os_packages.arch_pkgbuild import (
from pkgmgr.actions.install.installers.os_packages.debian_control import DebianControlInstaller # noqa: F401 ArchPkgbuildInstaller,
) # noqa: F401
from pkgmgr.actions.install.installers.os_packages.debian_control import (
DebianControlInstaller,
) # noqa: F401
from pkgmgr.actions.install.installers.os_packages.rpm_spec import RpmSpecInstaller # noqa: F401 from pkgmgr.actions.install.installers.os_packages.rpm_spec import RpmSpecInstaller # noqa: F401

View File

@@ -41,7 +41,9 @@ class BaseInstaller(ABC):
return caps return caps
for matcher in CAPABILITY_MATCHERS: for matcher in CAPABILITY_MATCHERS:
if matcher.applies_to_layer(self.layer) and matcher.is_provided(ctx, self.layer): if matcher.applies_to_layer(self.layer) and matcher.is_provided(
ctx, self.layer
):
caps.add(matcher.name) caps.add(matcher.name)
return caps return caps

View File

@@ -16,7 +16,9 @@ class MakefileInstaller(BaseInstaller):
def supports(self, ctx: RepoContext) -> bool: def supports(self, ctx: RepoContext) -> bool:
if os.environ.get("PKGMGR_DISABLE_MAKEFILE_INSTALLER") == "1": if os.environ.get("PKGMGR_DISABLE_MAKEFILE_INSTALLER") == "1":
if not ctx.quiet: if not ctx.quiet:
print("[INFO] PKGMGR_DISABLE_MAKEFILE_INSTALLER=1 skipping MakefileInstaller.") print(
"[INFO] PKGMGR_DISABLE_MAKEFILE_INSTALLER=1 skipping MakefileInstaller."
)
return False return False
makefile_path = os.path.join(ctx.repo_dir, self.MAKEFILE_NAME) makefile_path = os.path.join(ctx.repo_dir, self.MAKEFILE_NAME)
@@ -46,7 +48,9 @@ class MakefileInstaller(BaseInstaller):
return return
if not ctx.quiet: if not ctx.quiet:
print(f"[pkgmgr] Running make install for {ctx.identifier} (MakefileInstaller)") print(
f"[pkgmgr] Running make install for {ctx.identifier} (MakefileInstaller)"
)
run_command("make install", cwd=ctx.repo_dir, preview=ctx.preview) run_command("make install", cwd=ctx.repo_dir, preview=ctx.preview)

View File

@@ -57,7 +57,9 @@ class NixConflictResolver:
# 3) Fallback: output-name based lookup (also covers nix suggesting: `nix profile remove pkgmgr`) # 3) Fallback: output-name based lookup (also covers nix suggesting: `nix profile remove pkgmgr`)
if not tokens: if not tokens:
tokens = self._profile.find_remove_tokens_for_output(ctx, self._runner, output) tokens = self._profile.find_remove_tokens_for_output(
ctx, self._runner, output
)
if tokens: if tokens:
if not quiet: if not quiet:
@@ -94,7 +96,9 @@ class NixConflictResolver:
continue continue
if not quiet: if not quiet:
print("[nix] conflict detected but could not resolve profile entries to remove.") print(
"[nix] conflict detected but could not resolve profile entries to remove."
)
return False return False
return False return False

View File

@@ -75,7 +75,9 @@ class NixFlakeInstaller(BaseInstaller):
# Core install path # Core install path
# --------------------------------------------------------------------- # ---------------------------------------------------------------------
def _install_only(self, ctx: "RepoContext", output: str, allow_failure: bool) -> None: def _install_only(
self, ctx: "RepoContext", output: str, allow_failure: bool
) -> None:
install_cmd = f"nix profile install {self._installable(ctx, output)}" install_cmd = f"nix profile install {self._installable(ctx, output)}"
if not ctx.quiet: if not ctx.quiet:
@@ -96,7 +98,9 @@ class NixFlakeInstaller(BaseInstaller):
output=output, output=output,
): ):
if not ctx.quiet: if not ctx.quiet:
print(f"[nix] output '{output}' successfully installed after conflict cleanup.") print(
f"[nix] output '{output}' successfully installed after conflict cleanup."
)
return return
if not ctx.quiet: if not ctx.quiet:
@@ -107,20 +111,26 @@ class NixFlakeInstaller(BaseInstaller):
# If indices are supported, try legacy index-upgrade path. # If indices are supported, try legacy index-upgrade path.
if self._indices_supported is not False: if self._indices_supported is not False:
indices = self._profile.find_installed_indices_for_output(ctx, self._runner, output) indices = self._profile.find_installed_indices_for_output(
ctx, self._runner, output
)
upgraded = False upgraded = False
for idx in indices: for idx in indices:
if self._upgrade_index(ctx, idx): if self._upgrade_index(ctx, idx):
upgraded = True upgraded = True
if not ctx.quiet: if not ctx.quiet:
print(f"[nix] output '{output}' successfully upgraded (index {idx}).") print(
f"[nix] output '{output}' successfully upgraded (index {idx})."
)
if upgraded: if upgraded:
return return
if indices and not ctx.quiet: if indices and not ctx.quiet:
print(f"[nix] upgrade failed; removing indices {indices} and reinstalling '{output}'.") print(
f"[nix] upgrade failed; removing indices {indices} and reinstalling '{output}'."
)
for idx in indices: for idx in indices:
self._remove_index(ctx, idx) self._remove_index(ctx, idx)
@@ -139,7 +149,9 @@ class NixFlakeInstaller(BaseInstaller):
print(f"[nix] output '{output}' successfully re-installed.") print(f"[nix] output '{output}' successfully re-installed.")
return return
print(f"[ERROR] Failed to install Nix flake output '{output}' (exit {final.returncode})") print(
f"[ERROR] Failed to install Nix flake output '{output}' (exit {final.returncode})"
)
if not allow_failure: if not allow_failure:
raise SystemExit(final.returncode) raise SystemExit(final.returncode)
@@ -149,7 +161,9 @@ class NixFlakeInstaller(BaseInstaller):
# force_update path # force_update path
# --------------------------------------------------------------------- # ---------------------------------------------------------------------
def _force_upgrade_output(self, ctx: "RepoContext", output: str, allow_failure: bool) -> None: def _force_upgrade_output(
self, ctx: "RepoContext", output: str, allow_failure: bool
) -> None:
# Prefer token path if indices unsupported (new nix) # Prefer token path if indices unsupported (new nix)
if self._indices_supported is False: if self._indices_supported is False:
self._remove_tokens_for_output(ctx, output) self._remove_tokens_for_output(ctx, output)
@@ -158,14 +172,18 @@ class NixFlakeInstaller(BaseInstaller):
print(f"[nix] output '{output}' successfully upgraded.") print(f"[nix] output '{output}' successfully upgraded.")
return return
indices = self._profile.find_installed_indices_for_output(ctx, self._runner, output) indices = self._profile.find_installed_indices_for_output(
ctx, self._runner, output
)
upgraded_any = False upgraded_any = False
for idx in indices: for idx in indices:
if self._upgrade_index(ctx, idx): if self._upgrade_index(ctx, idx):
upgraded_any = True upgraded_any = True
if not ctx.quiet: if not ctx.quiet:
print(f"[nix] output '{output}' successfully upgraded (index {idx}).") print(
f"[nix] output '{output}' successfully upgraded (index {idx})."
)
if upgraded_any: if upgraded_any:
if not ctx.quiet: if not ctx.quiet:
@@ -173,7 +191,9 @@ class NixFlakeInstaller(BaseInstaller):
return return
if indices and not ctx.quiet: if indices and not ctx.quiet:
print(f"[nix] upgrade failed; removing indices {indices} and reinstalling '{output}'.") print(
f"[nix] upgrade failed; removing indices {indices} and reinstalling '{output}'."
)
for idx in indices: for idx in indices:
self._remove_index(ctx, idx) self._remove_index(ctx, idx)
@@ -223,7 +243,9 @@ class NixFlakeInstaller(BaseInstaller):
return return
if not ctx.quiet: if not ctx.quiet:
print(f"[nix] indices unsupported; removing by token(s): {', '.join(tokens)}") print(
f"[nix] indices unsupported; removing by token(s): {', '.join(tokens)}"
)
for t in tokens: for t in tokens:
self._runner.run(ctx, f"nix profile remove {t}", allow_failure=True) self._runner.run(ctx, f"nix profile remove {t}", allow_failure=True)

View File

@@ -101,7 +101,9 @@ class NixProfileInspector:
data = self.list_json(ctx, runner) data = self.list_json(ctx, runner)
entries = normalize_elements(data) entries = normalize_elements(data)
tokens: List[str] = [out] # critical: matches nix's own suggestion for conflicts tokens: List[str] = [
out
] # critical: matches nix's own suggestion for conflicts
for e in entries: for e in entries:
if entry_matches_output(e, out): if entry_matches_output(e, out):

View File

@@ -48,7 +48,9 @@ class NixProfileListReader:
return uniq return uniq
def indices_matching_store_prefixes(self, ctx: "RepoContext", prefixes: List[str]) -> List[int]: def indices_matching_store_prefixes(
self, ctx: "RepoContext", prefixes: List[str]
) -> List[int]:
prefixes = [self._store_prefix(p) for p in prefixes if p] prefixes = [self._store_prefix(p) for p in prefixes if p]
prefixes = [p for p in prefixes if p] prefixes = [p for p in prefixes if p]
if not prefixes: if not prefixes:

View File

@@ -11,6 +11,7 @@ if TYPE_CHECKING:
from pkgmgr.actions.install.context import RepoContext from pkgmgr.actions.install.context import RepoContext
from .runner import CommandRunner from .runner import CommandRunner
@dataclass(frozen=True) @dataclass(frozen=True)
class RetryPolicy: class RetryPolicy:
max_attempts: int = 7 max_attempts: int = 7
@@ -35,13 +36,19 @@ class GitHubRateLimitRetry:
install_cmd: str, install_cmd: str,
) -> RunResult: ) -> RunResult:
quiet = bool(getattr(ctx, "quiet", False)) quiet = bool(getattr(ctx, "quiet", False))
delays = list(self._fibonacci_backoff(self._policy.base_delay_seconds, self._policy.max_attempts)) delays = list(
self._fibonacci_backoff(
self._policy.base_delay_seconds, self._policy.max_attempts
)
)
last: RunResult | None = None last: RunResult | None = None
for attempt, base_delay in enumerate(delays, start=1): for attempt, base_delay in enumerate(delays, start=1):
if not quiet: if not quiet:
print(f"[nix] attempt {attempt}/{self._policy.max_attempts}: {install_cmd}") print(
f"[nix] attempt {attempt}/{self._policy.max_attempts}: {install_cmd}"
)
res = runner.run(ctx, install_cmd, allow_failure=True) res = runner.run(ctx, install_cmd, allow_failure=True)
last = res last = res
@@ -56,7 +63,9 @@ class GitHubRateLimitRetry:
if attempt >= self._policy.max_attempts: if attempt >= self._policy.max_attempts:
break break
jitter = random.randint(self._policy.jitter_seconds_min, self._policy.jitter_seconds_max) jitter = random.randint(
self._policy.jitter_seconds_min, self._policy.jitter_seconds_max
)
wait_time = base_delay + jitter wait_time = base_delay + jitter
if not quiet: if not quiet:
@@ -67,7 +76,11 @@ class GitHubRateLimitRetry:
time.sleep(wait_time) time.sleep(wait_time)
return last if last is not None else RunResult(returncode=1, stdout="", stderr="nix install retry failed") return (
last
if last is not None
else RunResult(returncode=1, stdout="", stderr="nix install retry failed")
)
@staticmethod @staticmethod
def _is_github_rate_limit_error(text: str) -> bool: def _is_github_rate_limit_error(text: str) -> bool:

View File

@@ -9,6 +9,7 @@ from .types import RunResult
if TYPE_CHECKING: if TYPE_CHECKING:
from pkgmgr.actions.install.context import RepoContext from pkgmgr.actions.install.context import RepoContext
class CommandRunner: class CommandRunner:
""" """
Executes commands (shell=True) inside a repository directory (if provided). Executes commands (shell=True) inside a repository directory (if provided).
@@ -40,7 +41,9 @@ class CommandRunner:
raise raise
return RunResult(returncode=1, stdout="", stderr=str(e)) return RunResult(returncode=1, stdout="", stderr=str(e))
res = RunResult(returncode=p.returncode, stdout=p.stdout or "", stderr=p.stderr or "") res = RunResult(
returncode=p.returncode, stdout=p.stdout or "", stderr=p.stderr or ""
)
if res.returncode != 0 and not quiet: if res.returncode != 0 and not quiet:
self._print_compact_failure(res) self._print_compact_failure(res)

View File

@@ -20,7 +20,9 @@ class NixConflictTextParser:
tokens: List[str] = [] tokens: List[str] = []
for m in pat.finditer(text or ""): for m in pat.finditer(text or ""):
t = (m.group(1) or "").strip() t = (m.group(1) or "").strip()
if (t.startswith("'") and t.endswith("'")) or (t.startswith('"') and t.endswith('"')): if (t.startswith("'") and t.endswith("'")) or (
t.startswith('"') and t.endswith('"')
):
t = t[1:-1] t = t[1:-1]
if t: if t:
tokens.append(t) tokens.append(t)

View File

@@ -14,7 +14,9 @@ class PythonInstaller(BaseInstaller):
def supports(self, ctx: RepoContext) -> bool: def supports(self, ctx: RepoContext) -> bool:
if os.environ.get("PKGMGR_DISABLE_PYTHON_INSTALLER") == "1": if os.environ.get("PKGMGR_DISABLE_PYTHON_INSTALLER") == "1":
print("[INFO] PythonInstaller disabled via PKGMGR_DISABLE_PYTHON_INSTALLER.") print(
"[INFO] PythonInstaller disabled via PKGMGR_DISABLE_PYTHON_INSTALLER."
)
return False return False
return os.path.exists(os.path.join(ctx.repo_dir, "pyproject.toml")) return os.path.exists(os.path.join(ctx.repo_dir, "pyproject.toml"))

View File

@@ -132,7 +132,11 @@ class InstallationPipeline:
continue continue
if not quiet: if not quiet:
if ctx.force_update and state.layer is not None and installer_layer == state.layer: if (
ctx.force_update
and state.layer is not None
and installer_layer == state.layer
):
print( print(
f"[pkgmgr] Running installer {installer.__class__.__name__} " f"[pkgmgr] Running installer {installer.__class__.__name__} "
f"for {identifier} in '{repo_dir}' (upgrade requested)..." f"for {identifier} in '{repo_dir}' (upgrade requested)..."

View File

@@ -3,7 +3,7 @@ from __future__ import annotations
import os import os
from typing import Optional, Set from typing import Optional, Set
from pkgmgr.core.git.errors import GitError from pkgmgr.core.git.errors import GitRunError
from pkgmgr.core.git.commands import ( from pkgmgr.core.git.commands import (
GitAddRemoteError, GitAddRemoteError,
GitAddRemotePushUrlError, GitAddRemotePushUrlError,
@@ -90,7 +90,7 @@ def determine_primary_remote_url(
def has_origin_remote(repo_dir: str) -> bool: def has_origin_remote(repo_dir: str) -> bool:
try: try:
return "origin" in list_remotes(cwd=repo_dir) return "origin" in list_remotes(cwd=repo_dir)
except GitError: except GitRunError:
return False return False
@@ -122,7 +122,7 @@ def _ensure_additional_push_urls(
try: try:
existing = get_remote_push_urls("origin", cwd=repo_dir) existing = get_remote_push_urls("origin", cwd=repo_dir)
except GitError: except GitRunError:
existing = set() existing = set()
for url in sorted(desired - existing): for url in sorted(desired - existing):

View File

@@ -1,8 +1,9 @@
from __future__ import annotations from __future__ import annotations
import os import os
from collections.abc import Iterable, Mapping
from typing import Union
from urllib.parse import urlparse from urllib.parse import urlparse
from typing import Mapping
from .types import MirrorMap, Repository from .types import MirrorMap, Repository
@@ -32,7 +33,7 @@ def read_mirrors_file(repo_dir: str, filename: str = "MIRRORS") -> MirrorMap:
""" """
Supports: Supports:
NAME URL NAME URL
URL auto name = hostname URL -> auto-generate name from hostname
""" """
path = os.path.join(repo_dir, filename) path = os.path.join(repo_dir, filename)
mirrors: MirrorMap = {} mirrors: MirrorMap = {}
@@ -52,7 +53,8 @@ def read_mirrors_file(repo_dir: str, filename: str = "MIRRORS") -> MirrorMap:
# Case 1: "name url" # Case 1: "name url"
if len(parts) == 2: if len(parts) == 2:
name, url = parts name, url = parts
# Case 2: "url" → auto-generate name
# Case 2: "url" -> auto name
elif len(parts) == 1: elif len(parts) == 1:
url = parts[0] url = parts[0]
parsed = urlparse(url) parsed = urlparse(url)
@@ -67,21 +69,56 @@ def read_mirrors_file(repo_dir: str, filename: str = "MIRRORS") -> MirrorMap:
continue continue
mirrors[name] = url mirrors[name] = url
except OSError as exc: except OSError as exc:
print(f"[WARN] Could not read MIRRORS file at {path}: {exc}") print(f"[WARN] Could not read MIRRORS file at {path}: {exc}")
return mirrors return mirrors
MirrorsInput = Union[Mapping[str, str], Iterable[str]]
def write_mirrors_file( def write_mirrors_file(
repo_dir: str, repo_dir: str,
mirrors: Mapping[str, str], mirrors: MirrorsInput,
filename: str = "MIRRORS", filename: str = "MIRRORS",
preview: bool = False, preview: bool = False,
) -> None: ) -> None:
"""
Write MIRRORS in one of two formats:
1) Mapping[str, str] -> "NAME URL" per line (legacy / compatible)
2) Iterable[str] -> "URL" per line (new preferred)
Strings are treated as a single URL (not iterated character-by-character).
"""
path = os.path.join(repo_dir, filename) path = os.path.join(repo_dir, filename)
lines = [f"{name} {url}" for name, url in sorted(mirrors.items())]
lines: list[str]
if isinstance(mirrors, Mapping):
items = [
(str(name), str(url))
for name, url in mirrors.items()
if url is not None and str(url).strip()
]
items.sort(key=lambda x: (x[0], x[1]))
lines = [f"{name} {url}" for name, url in items]
else:
if isinstance(mirrors, (str, bytes)):
urls = [str(mirrors).strip()]
else:
urls = [
str(url).strip()
for url in mirrors
if url is not None and str(url).strip()
]
urls = sorted(set(urls))
lines = urls
content = "\n".join(lines) + ("\n" if lines else "") content = "\n".join(lines) + ("\n" if lines else "")
if preview: if preview:
@@ -94,5 +131,6 @@ def write_mirrors_file(
with open(path, "w", encoding="utf-8") as fh: with open(path, "w", encoding="utf-8") as fh:
fh.write(content) fh.write(content)
print(f"[INFO] Wrote MIRRORS file at {path}") print(f"[INFO] Wrote MIRRORS file at {path}")
except OSError as exc: except OSError as exc:
print(f"[ERROR] Failed to write MIRRORS file at {path}: {exc}") print(f"[ERROR] Failed to write MIRRORS file at {path}: {exc}")

View File

@@ -16,6 +16,7 @@ from .types import MirrorMap, Repository
# Helpers # Helpers
# ----------------------------------------------------------------------------- # -----------------------------------------------------------------------------
def _repo_key(repo: Repository) -> Tuple[str, str, str]: def _repo_key(repo: Repository) -> Tuple[str, str, str]:
""" """
Normalised key for identifying a repository in config files. Normalised key for identifying a repository in config files.
@@ -47,6 +48,7 @@ def _load_user_config(path: str) -> Dict[str, object]:
# Main merge command # Main merge command
# ----------------------------------------------------------------------------- # -----------------------------------------------------------------------------
def merge_mirrors( def merge_mirrors(
selected_repos: List[Repository], selected_repos: List[Repository],
repositories_base_dir: str, repositories_base_dir: str,

View File

@@ -66,7 +66,9 @@ def _setup_remote_mirrors_for_repo(
# Probe only git URLs (do not try ls-remote against PyPI etc.) # Probe only git URLs (do not try ls-remote against PyPI etc.)
# If there are no mirrors at all, probe the primary git URL. # If there are no mirrors at all, probe the primary git URL.
git_mirrors = {k: v for k, v in ctx.resolved_mirrors.items() if _is_git_remote_url(v)} git_mirrors = {
k: v for k, v in ctx.resolved_mirrors.items() if _is_git_remote_url(v)
}
if not git_mirrors: if not git_mirrors:
primary = determine_primary_remote_url(repo, ctx) primary = determine_primary_remote_url(repo, ctx)

View File

@@ -4,7 +4,16 @@ from pkgmgr.core.repository.dir import get_repo_dir
from pkgmgr.core.command.run import run_command from pkgmgr.core.command.run import run_command
import sys import sys
def exec_proxy_command(proxy_prefix: str, selected_repos, repositories_base_dir, all_repos, proxy_command: str, extra_args, preview: bool):
def exec_proxy_command(
proxy_prefix: str,
selected_repos,
repositories_base_dir,
all_repos,
proxy_command: str,
extra_args,
preview: bool,
):
"""Execute a given proxy command with extra arguments for each repository.""" """Execute a given proxy command with extra arguments for each repository."""
error_repos = [] error_repos = []
max_exit_code = 0 max_exit_code = 0
@@ -22,7 +31,9 @@ def exec_proxy_command(proxy_prefix: str, selected_repos, repositories_base_dir,
try: try:
run_command(full_cmd, cwd=repo_dir, preview=preview) run_command(full_cmd, cwd=repo_dir, preview=preview)
except SystemExit as e: except SystemExit as e:
print(f"[ERROR] Command failed in {repo_identifier} with exit code {e.code}.") print(
f"[ERROR] Command failed in {repo_identifier} with exit code {e.code}."
)
error_repos.append((repo_identifier, e.code)) error_repos.append((repo_identifier, e.code))
max_exit_code = max(max_exit_code, e.code) max_exit_code = max(max_exit_code, e.code)

View File

@@ -121,7 +121,7 @@ def update_pyproject_version(
pattern = r'^(version\s*=\s*")([^"]+)(")' pattern = r'^(version\s*=\s*")([^"]+)(")'
new_content, count = re.subn( new_content, count = re.subn(
pattern, pattern,
lambda m: f'{m.group(1)}{new_version}{m.group(3)}', lambda m: f"{m.group(1)}{new_version}{m.group(3)}",
content, content,
flags=re.MULTILINE, flags=re.MULTILINE,
) )
@@ -162,7 +162,7 @@ def update_flake_version(
pattern = r'(version\s*=\s*")([^"]+)(")' pattern = r'(version\s*=\s*")([^"]+)(")'
new_content, count = re.subn( new_content, count = re.subn(
pattern, pattern,
lambda m: f'{m.group(1)}{new_version}{m.group(3)}', lambda m: f"{m.group(1)}{new_version}{m.group(3)}",
content, content,
) )

View File

@@ -80,7 +80,9 @@ def is_highest_version_tag(tag: str) -> bool:
return True return True
latest = max(parsed_all) latest = max(parsed_all)
print(f"[INFO] Latest tag (parsed): v{'.'.join(map(str, latest))}, Current tag: {tag}") print(
f"[INFO] Latest tag (parsed): v{'.'.join(map(str, latest))}, Current tag: {tag}"
)
return parsed_current >= latest return parsed_current >= latest
@@ -93,7 +95,9 @@ def update_latest_tag(new_tag: str, *, preview: bool = False) -> None:
- 'latest' is forced (floating tag), therefore the push uses --force. - 'latest' is forced (floating tag), therefore the push uses --force.
""" """
target_ref = f"{new_tag}^{{}}" target_ref = f"{new_tag}^{{}}"
print(f"[INFO] Updating 'latest' tag to point at {new_tag} (commit {target_ref})...") print(
f"[INFO] Updating 'latest' tag to point at {new_tag} (commit {target_ref})..."
)
tag_force_annotated( tag_force_annotated(
name="latest", name="latest",

View File

@@ -5,7 +5,7 @@ import sys
from typing import Optional from typing import Optional
from pkgmgr.actions.branch import close_branch from pkgmgr.actions.branch import close_branch
from pkgmgr.core.git import GitError from pkgmgr.core.git import GitRunError
from pkgmgr.core.git.commands import add, commit, push, tag_annotated from pkgmgr.core.git.commands import add, commit, push, tag_annotated
from pkgmgr.core.git.queries import get_current_branch from pkgmgr.core.git.queries import get_current_branch
from pkgmgr.core.repository.paths import resolve_repo_paths from pkgmgr.core.repository.paths import resolve_repo_paths
@@ -40,7 +40,7 @@ def _release_impl(
# Determine current branch early # Determine current branch early
try: try:
branch = get_current_branch() or "main" branch = get_current_branch() or "main"
except GitError: except GitRunError:
branch = "main" branch = "main"
print(f"Releasing on branch: {branch}") print(f"Releasing on branch: {branch}")
@@ -76,7 +76,9 @@ def _release_impl(
if paths.arch_pkgbuild: if paths.arch_pkgbuild:
update_pkgbuild_version(paths.arch_pkgbuild, new_ver_str, preview=preview) update_pkgbuild_version(paths.arch_pkgbuild, new_ver_str, preview=preview)
else: else:
print("[INFO] No PKGBUILD found (packaging/arch/PKGBUILD or PKGBUILD). Skipping.") print(
"[INFO] No PKGBUILD found (packaging/arch/PKGBUILD or PKGBUILD). Skipping."
)
if paths.rpm_spec: if paths.rpm_spec:
update_spec_version(paths.rpm_spec, new_ver_str, preview=preview) update_spec_version(paths.rpm_spec, new_ver_str, preview=preview)
@@ -123,7 +125,9 @@ def _release_impl(
paths.rpm_spec, paths.rpm_spec,
paths.debian_changelog, paths.debian_changelog,
] ]
existing_files = [p for p in files_to_add if isinstance(p, str) and p and os.path.exists(p)] existing_files = [
p for p in files_to_add if isinstance(p, str) and p and os.path.exists(p)
]
if preview: if preview:
add(existing_files, preview=True) add(existing_files, preview=True)
@@ -135,13 +139,17 @@ def _release_impl(
if is_highest_version_tag(new_tag): if is_highest_version_tag(new_tag):
update_latest_tag(new_tag, preview=True) update_latest_tag(new_tag, preview=True)
else: else:
print(f"[PREVIEW] Skipping 'latest' update (tag {new_tag} is not the highest).") print(
f"[PREVIEW] Skipping 'latest' update (tag {new_tag} is not the highest)."
)
if close and branch not in ("main", "master"): if close and branch not in ("main", "master"):
if force: if force:
print(f"[PREVIEW] Would delete branch {branch} (forced).") print(f"[PREVIEW] Would delete branch {branch} (forced).")
else: else:
print(f"[PREVIEW] Would ask whether to delete branch {branch} after release.") print(
f"[PREVIEW] Would ask whether to delete branch {branch} after release."
)
return return
add(existing_files, preview=False) add(existing_files, preview=False)
@@ -157,8 +165,10 @@ def _release_impl(
if is_highest_version_tag(new_tag): if is_highest_version_tag(new_tag):
update_latest_tag(new_tag, preview=False) update_latest_tag(new_tag, preview=False)
else: else:
print(f"[INFO] Skipping 'latest' update (tag {new_tag} is not the highest).") print(
except GitError as exc: f"[INFO] Skipping 'latest' update (tag {new_tag} is not the highest)."
)
except GitRunError as exc:
print(f"[WARN] Failed to update floating 'latest' tag for {new_tag}: {exc}") print(f"[WARN] Failed to update floating 'latest' tag for {new_tag}: {exc}")
print("'latest' tag was not updated.") print("'latest' tag was not updated.")
@@ -166,7 +176,9 @@ def _release_impl(
if close: if close:
if branch in ("main", "master"): if branch in ("main", "master"):
print(f"[INFO] close=True but current branch is {branch}; skipping branch deletion.") print(
f"[INFO] close=True but current branch is {branch}; skipping branch deletion."
)
return return
if not should_delete_branch(force=force): if not should_delete_branch(force=force):

View File

@@ -55,7 +55,9 @@ def clone_repos(
clone_url = _build_clone_url(repo, clone_mode) clone_url = _build_clone_url(repo, clone_mode)
if not clone_url: if not clone_url:
print(f"[WARNING] Cannot build clone URL for '{repo_identifier}'. Skipping.") print(
f"[WARNING] Cannot build clone URL for '{repo_identifier}'. Skipping."
)
continue continue
shallow = clone_mode == "shallow" shallow = clone_mode == "shallow"
@@ -84,7 +86,11 @@ def clone_repos(
continue continue
print(f"[WARNING] SSH clone failed for '{repo_identifier}': {exc}") print(f"[WARNING] SSH clone failed for '{repo_identifier}': {exc}")
choice = input("Do you want to attempt HTTPS clone instead? (y/N): ").strip().lower() choice = (
input("Do you want to attempt HTTPS clone instead? (y/N): ")
.strip()
.lower()
)
if choice != "y": if choice != "y":
print(f"[INFO] HTTPS clone not attempted for '{repo_identifier}'.") print(f"[INFO] HTTPS clone not attempted for '{repo_identifier}'.")
continue continue

View File

@@ -12,8 +12,8 @@ class MirrorBootstrapper:
""" """
MIRRORS is the single source of truth. MIRRORS is the single source of truth.
We write defaults to MIRRORS and then call mirror setup which will Defaults are written to MIRRORS and mirror setup derives
configure git remotes based on MIRRORS content (but only for git URLs). git remotes exclusively from that file (git URLs only).
""" """
def write_defaults( def write_defaults(
@@ -25,10 +25,8 @@ class MirrorBootstrapper:
preview: bool, preview: bool,
) -> None: ) -> None:
mirrors = { mirrors = {
# preferred SSH url is supplied by CreateRepoPlanner.primary_remote primary,
"origin": primary, f"https://pypi.org/project/{name}/",
# metadata only: must NEVER be configured as a git remote
"pypi": f"https://pypi.org/project/{name}/",
} }
write_mirrors_file(repo_dir, mirrors, preview=preview) write_mirrors_file(repo_dir, mirrors, preview=preview)
@@ -41,7 +39,8 @@ class MirrorBootstrapper:
preview: bool, preview: bool,
remote: bool, remote: bool,
) -> None: ) -> None:
# IMPORTANT: do NOT set repo["mirrors"] here. # IMPORTANT:
# Do NOT set repo["mirrors"] here.
# MIRRORS file is the single source of truth. # MIRRORS file is the single source of truth.
setup_mirrors( setup_mirrors(
selected_repos=[repo], selected_repos=[repo],

View File

@@ -63,6 +63,4 @@ def _strip_git_suffix(name: str) -> str:
def _ensure_valid_repo_name(name: str) -> None: def _ensure_valid_repo_name(name: str) -> None:
if not _NAME_RE.fullmatch(name): if not _NAME_RE.fullmatch(name):
raise ValueError( raise ValueError("Repository name must match: lowercase a-z, 0-9, '_' and '-'.")
"Repository name must match: lowercase a-z, 0-9, '_' and '-'."
)

View File

@@ -66,9 +66,7 @@ class TemplateRenderer:
for root, _, files in os.walk(self.templates_dir): for root, _, files in os.walk(self.templates_dir):
for fn in files: for fn in files:
if fn.endswith(".j2"): if fn.endswith(".j2"):
rel = os.path.relpath( rel = os.path.relpath(os.path.join(root, fn), self.templates_dir)
os.path.join(root, fn), self.templates_dir
)
print(f"[Preview] Would render template: {rel} -> {rel[:-3]}") print(f"[Preview] Would render template: {rel} -> {rel[:-3]}")
@staticmethod @staticmethod

View File

@@ -24,9 +24,13 @@ def deinstall_repos(
# Remove alias link/file (interactive) # Remove alias link/file (interactive)
if os.path.exists(alias_path): if os.path.exists(alias_path):
confirm = input( confirm = (
input(
f"Are you sure you want to delete link '{alias_path}' for {repo_identifier}? [y/N]: " f"Are you sure you want to delete link '{alias_path}' for {repo_identifier}? [y/N]: "
).strip().lower() )
.strip()
.lower()
)
if confirm == "y": if confirm == "y":
if preview: if preview:
print(f"[Preview] Would remove link '{alias_path}'.") print(f"[Preview] Would remove link '{alias_path}'.")

View File

@@ -3,19 +3,30 @@ import os
from pkgmgr.core.repository.identifier import get_repo_identifier from pkgmgr.core.repository.identifier import get_repo_identifier
from pkgmgr.core.repository.dir import get_repo_dir from pkgmgr.core.repository.dir import get_repo_dir
def delete_repos(selected_repos, repositories_base_dir, all_repos, preview=False): def delete_repos(selected_repos, repositories_base_dir, all_repos, preview=False):
for repo in selected_repos: for repo in selected_repos:
repo_identifier = get_repo_identifier(repo, all_repos) repo_identifier = get_repo_identifier(repo, all_repos)
repo_dir = get_repo_dir(repositories_base_dir, repo) repo_dir = get_repo_dir(repositories_base_dir, repo)
if os.path.exists(repo_dir): if os.path.exists(repo_dir):
confirm = input(f"Are you sure you want to delete directory '{repo_dir}' for {repo_identifier}? [y/N]: ").strip().lower() confirm = (
input(
f"Are you sure you want to delete directory '{repo_dir}' for {repo_identifier}? [y/N]: "
)
.strip()
.lower()
)
if confirm == "y": if confirm == "y":
if preview: if preview:
print(f"[Preview] Would delete directory '{repo_dir}' for {repo_identifier}.") print(
f"[Preview] Would delete directory '{repo_dir}' for {repo_identifier}."
)
else: else:
try: try:
shutil.rmtree(repo_dir) shutil.rmtree(repo_dir)
print(f"Deleted repository directory '{repo_dir}' for {repo_identifier}.") print(
f"Deleted repository directory '{repo_dir}' for {repo_identifier}."
)
except Exception as e: except Exception as e:
print(f"Error deleting '{repo_dir}' for {repo_identifier}: {e}") print(f"Error deleting '{repo_dir}' for {repo_identifier}: {e}")
else: else:

View File

@@ -233,9 +233,7 @@ def list_repositories(
categories.append(str(repo["category"])) categories.append(str(repo["category"]))
yaml_tags: List[str] = list(map(str, repo.get("tags", []))) yaml_tags: List[str] = list(map(str, repo.get("tags", [])))
display_tags: List[str] = sorted( display_tags: List[str] = sorted(set(yaml_tags + list(map(str, extra_tags))))
set(yaml_tags + list(map(str, extra_tags)))
)
rows.append( rows.append(
{ {
@@ -288,13 +286,7 @@ def list_repositories(
status_padded = status.ljust(status_width) status_padded = status.ljust(status_width)
status_colored = _color_status(status_padded) status_colored = _color_status(status_padded)
print( print(f"{ident_col} {status_colored} {cat_col} {tag_col} {dir_col}")
f"{ident_col} "
f"{status_colored} "
f"{cat_col} "
f"{tag_col} "
f"{dir_col}"
)
# ------------------------------------------------------------------ # ------------------------------------------------------------------
# Detailed section (alias value red, same status coloring) # Detailed section (alias value red, same status coloring)

View File

@@ -55,12 +55,16 @@ class UpdateManager:
code = exc.code if isinstance(exc.code, int) else str(exc.code) code = exc.code if isinstance(exc.code, int) else str(exc.code)
failures.append((identifier, f"pull failed (exit={code})")) failures.append((identifier, f"pull failed (exit={code})"))
if not quiet: if not quiet:
print(f"[Warning] update: pull failed for {identifier} (exit={code}). Continuing...") print(
f"[Warning] update: pull failed for {identifier} (exit={code}). Continuing..."
)
continue continue
except Exception as exc: except Exception as exc:
failures.append((identifier, f"pull failed: {exc}")) failures.append((identifier, f"pull failed: {exc}"))
if not quiet: if not quiet:
print(f"[Warning] update: pull failed for {identifier}: {exc}. Continuing...") print(
f"[Warning] update: pull failed for {identifier}: {exc}. Continuing..."
)
continue continue
try: try:
@@ -82,12 +86,16 @@ class UpdateManager:
code = exc.code if isinstance(exc.code, int) else str(exc.code) code = exc.code if isinstance(exc.code, int) else str(exc.code)
failures.append((identifier, f"install failed (exit={code})")) failures.append((identifier, f"install failed (exit={code})"))
if not quiet: if not quiet:
print(f"[Warning] update: install failed for {identifier} (exit={code}). Continuing...") print(
f"[Warning] update: install failed for {identifier} (exit={code}). Continuing..."
)
continue continue
except Exception as exc: except Exception as exc:
failures.append((identifier, f"install failed: {exc}")) failures.append((identifier, f"install failed: {exc}"))
if not quiet: if not quiet:
print(f"[Warning] update: install failed for {identifier}: {exc}. Continuing...") print(
f"[Warning] update: install failed for {identifier}: {exc}. Continuing..."
)
continue continue
if failures and not quiet: if failures and not quiet:

View File

@@ -31,6 +31,7 @@ class OSReleaseInfo:
""" """
Minimal /etc/os-release representation for distro detection. Minimal /etc/os-release representation for distro detection.
""" """
id: str = "" id: str = ""
id_like: str = "" id_like: str = ""
pretty_name: str = "" pretty_name: str = ""
@@ -63,4 +64,6 @@ class OSReleaseInfo:
def is_fedora_family(self) -> bool: def is_fedora_family(self) -> bool:
ids = self.ids() ids = self.ids()
return bool(ids.intersection({"fedora", "rhel", "centos", "rocky", "almalinux"})) return bool(
ids.intersection({"fedora", "rhel", "centos", "rocky", "almalinux"})
)

View File

@@ -58,7 +58,9 @@ class SystemUpdater:
run_command("sudo pacman -Syu --noconfirm", preview=preview) run_command("sudo pacman -Syu --noconfirm", preview=preview)
return return
print("[Warning] Cannot update Arch system: missing required tools (sudo/yay/pacman).") print(
"[Warning] Cannot update Arch system: missing required tools (sudo/yay/pacman)."
)
def _update_debian(self, *, preview: bool) -> None: def _update_debian(self, *, preview: bool) -> None:
from pkgmgr.core.command.run import run_command from pkgmgr.core.command.run import run_command
@@ -67,7 +69,9 @@ class SystemUpdater:
apt_get = shutil.which("apt-get") apt_get = shutil.which("apt-get")
if not (sudo and apt_get): if not (sudo and apt_get):
print("[Warning] Cannot update Debian/Ubuntu system: missing required tools (sudo/apt-get).") print(
"[Warning] Cannot update Debian/Ubuntu system: missing required tools (sudo/apt-get)."
)
return return
env = "DEBIAN_FRONTEND=noninteractive" env = "DEBIAN_FRONTEND=noninteractive"

View File

@@ -29,6 +29,7 @@ For details on any command, run:
\033[1mpkgmgr <command> --help\033[0m \033[1mpkgmgr <command> --help\033[0m
""" """
def main() -> None: def main() -> None:
""" """
Entry point for the pkgmgr CLI. Entry point for the pkgmgr CLI.
@@ -41,9 +42,7 @@ def main() -> None:
repositories_dir = os.path.expanduser( repositories_dir = os.path.expanduser(
directories.get("repositories", "~/Repositories") directories.get("repositories", "~/Repositories")
) )
binaries_dir = os.path.expanduser( binaries_dir = os.path.expanduser(directories.get("binaries", "~/.local/bin"))
directories.get("binaries", "~/.local/bin")
)
# Ensure the merged config actually contains the resolved directories # Ensure the merged config actually contains the resolved directories
config_merged.setdefault("directories", {}) config_merged.setdefault("directories", {})

View File

@@ -135,9 +135,7 @@ def handle_changelog(
target_tag=range_arg, target_tag=range_arg,
) )
if cur_tag is None: if cur_tag is None:
print( print(f"[WARN] Tag {range_arg!r} not found or not a SemVer tag.")
f"[WARN] Tag {range_arg!r} not found or not a SemVer tag."
)
print("[INFO] Falling back to full history.") print("[INFO] Falling back to full history.")
from_ref = None from_ref = None
to_ref = None to_ref = None

View File

@@ -213,9 +213,7 @@ def handle_config(args, ctx: CLIContext) -> None:
) )
if key == mod_key: if key == mod_key:
entry["ignore"] = args.set == "true" entry["ignore"] = args.set == "true"
print( print(f"Set ignore for {key} to {entry['ignore']}")
f"Set ignore for {key} to {entry['ignore']}"
)
save_user_config(user_config, user_config_path) save_user_config(user_config, user_config_path)
return return

View File

@@ -4,7 +4,12 @@ from __future__ import annotations
import sys import sys
from typing import Any, Dict, List from typing import Any, Dict, List
from pkgmgr.actions.mirror import diff_mirrors, list_mirrors, merge_mirrors, setup_mirrors from pkgmgr.actions.mirror import (
diff_mirrors,
list_mirrors,
merge_mirrors,
setup_mirrors,
)
from pkgmgr.cli.context import CLIContext from pkgmgr.cli.context import CLIContext
Repository = Dict[str, Any] Repository = Dict[str, Any]
@@ -56,11 +61,15 @@ def handle_mirror_command(
preview = getattr(args, "preview", False) preview = getattr(args, "preview", False)
if source == target: if source == target:
print("[ERROR] For 'mirror merge', source and target must differ (config vs file).") print(
"[ERROR] For 'mirror merge', source and target must differ (config vs file)."
)
sys.exit(2) sys.exit(2)
explicit_config_path = getattr(args, "config_path", None) explicit_config_path = getattr(args, "config_path", None)
user_config_path = explicit_config_path or getattr(ctx, "user_config_path", None) user_config_path = explicit_config_path or getattr(
ctx, "user_config_path", None
)
merge_mirrors( merge_mirrors(
selected_repos=selected, selected_repos=selected,

View File

@@ -18,7 +18,9 @@ def handle_publish(args, ctx: CLIContext, selected: List[Repository]) -> None:
for repo in selected: for repo in selected:
identifier = get_repo_identifier(repo, ctx.all_repositories) identifier = get_repo_identifier(repo, ctx.all_repositories)
repo_dir = repo.get("directory") or get_repo_dir(ctx.repositories_base_dir, repo) repo_dir = repo.get("directory") or get_repo_dir(
ctx.repositories_base_dir, repo
)
if not os.path.isdir(repo_dir): if not os.path.isdir(repo_dir):
print(f"[WARN] Skipping {identifier}: directory missing.") print(f"[WARN] Skipping {identifier}: directory missing.")

View File

@@ -36,9 +36,13 @@ def handle_release(
identifier = get_repo_identifier(repo, ctx.all_repositories) identifier = get_repo_identifier(repo, ctx.all_repositories)
try: try:
repo_dir = repo.get("directory") or get_repo_dir(ctx.repositories_base_dir, repo) repo_dir = repo.get("directory") or get_repo_dir(
ctx.repositories_base_dir, repo
)
except Exception as exc: except Exception as exc:
print(f"[WARN] Skipping repository {identifier}: failed to resolve directory: {exc}") print(
f"[WARN] Skipping repository {identifier}: failed to resolve directory: {exc}"
)
continue continue
if not os.path.isdir(repo_dir): if not os.path.isdir(repo_dir):

View File

@@ -32,9 +32,8 @@ def _resolve_repository_directory(repository: Repository, ctx: CLIContext) -> st
if repo_dir: if repo_dir:
return repo_dir return repo_dir
base_dir = ( base_dir = getattr(ctx, "repositories_base_dir", None) or getattr(
getattr(ctx, "repositories_base_dir", None) ctx, "repositories_dir", None
or getattr(ctx, "repositories_dir", None)
) )
if not base_dir: if not base_dir:
raise RuntimeError( raise RuntimeError(

View File

@@ -1,64 +1,27 @@
from __future__ import annotations from __future__ import annotations
import json
import os
from typing import Any, Dict, List from typing import Any, Dict, List
from pkgmgr.cli.context import CLIContext from pkgmgr.cli.context import CLIContext
from pkgmgr.cli.tools import open_vscode_workspace
from pkgmgr.cli.tools.paths import resolve_repository_path
from pkgmgr.core.command.run import run_command from pkgmgr.core.command.run import run_command
from pkgmgr .core .repository .identifier import get_repo_identifier
from pkgmgr .core .repository .dir import get_repo_dir
Repository = Dict[str, Any] Repository = Dict[str, Any]
def _resolve_repository_path(repository: Repository, ctx: CLIContext) -> str:
"""
Resolve the filesystem path for a repository.
Priority:
1. Use explicit keys if present (directory / path / workspace / workspace_dir).
2. Fallback to get_repo_dir(...) using the repositories base directory
from the CLI context.
"""
# 1) Explicit path-like keys on the repository object
for key in ("directory", "path", "workspace", "workspace_dir"):
value = repository.get(key)
if value:
return value
# 2) Fallback: compute from base dir + repository metadata
base_dir = (
getattr(ctx, "repositories_base_dir", None)
or getattr(ctx, "repositories_dir", None)
)
if not base_dir:
raise RuntimeError(
"Cannot resolve repositories base directory from context; "
"expected ctx.repositories_base_dir or ctx.repositories_dir."
)
return get_repo_dir(base_dir, repository)
def handle_tools_command( def handle_tools_command(
args, args,
ctx: CLIContext, ctx: CLIContext,
selected: List[Repository], selected: List[Repository],
) -> None: ) -> None:
# ------------------------------------------------------------------ # ------------------------------------------------------------------
# nautilus "explore" command # nautilus "explore" command
# ------------------------------------------------------------------ # ------------------------------------------------------------------
if args.command == "explore": if args.command == "explore":
for repository in selected: for repository in selected:
repo_path = _resolve_repository_path(repository, ctx) repo_path = resolve_repository_path(repository, ctx)
run_command( run_command(f'nautilus "{repo_path}" & disown')
f'nautilus "{repo_path}" & disown'
)
return return
# ------------------------------------------------------------------ # ------------------------------------------------------------------
@@ -66,50 +29,13 @@ def handle_tools_command(
# ------------------------------------------------------------------ # ------------------------------------------------------------------
if args.command == "terminal": if args.command == "terminal":
for repository in selected: for repository in selected:
repo_path = _resolve_repository_path(repository, ctx) repo_path = resolve_repository_path(repository, ctx)
run_command( run_command(f'gnome-terminal --tab --working-directory="{repo_path}"')
f'gnome-terminal --tab --working-directory="{repo_path}"'
)
return return
# ------------------------------------------------------------------ # ------------------------------------------------------------------
# VS Code workspace command # VS Code workspace command
# ------------------------------------------------------------------ # ------------------------------------------------------------------
if args.command == "code": if args.command == "code":
if not selected: open_vscode_workspace(ctx, selected)
print("No repositories selected.")
return
identifiers = [
get_repo_identifier(repo, ctx.all_repositories)
for repo in selected
]
sorted_identifiers = sorted(identifiers)
workspace_name = "_".join(sorted_identifiers) + ".code-workspace"
directories_cfg = ctx.config_merged.get("directories") or {}
workspaces_dir = os.path.expanduser(
directories_cfg.get("workspaces", "~/Workspaces")
)
os.makedirs(workspaces_dir, exist_ok=True)
workspace_file = os.path.join(workspaces_dir, workspace_name)
folders = [
{"path": _resolve_repository_path(repository, ctx)}
for repository in selected
]
workspace_data = {
"folders": folders,
"settings": {},
}
if not os.path.exists(workspace_file):
with open(workspace_file, "w", encoding="utf-8") as f:
json.dump(workspace_data, f, indent=4)
print(f"Created workspace file: {workspace_file}")
else:
print(f"Using existing workspace file: {workspace_file}")
run_command(f'code "{workspace_file}"')
return return

View File

@@ -33,8 +33,7 @@ def add_branch_subparsers(
"name", "name",
nargs="?", nargs="?",
help=( help=(
"Name of the new branch (optional; will be asked interactively " "Name of the new branch (optional; will be asked interactively if omitted)"
"if omitted)"
), ),
) )
branch_open.add_argument( branch_open.add_argument(
@@ -54,8 +53,7 @@ def add_branch_subparsers(
"name", "name",
nargs="?", nargs="?",
help=( help=(
"Name of the branch to close (optional; current branch is used " "Name of the branch to close (optional; current branch is used if omitted)"
"if omitted)"
), ),
) )
branch_close.add_argument( branch_close.add_argument(
@@ -84,8 +82,7 @@ def add_branch_subparsers(
"name", "name",
nargs="?", nargs="?",
help=( help=(
"Name of the branch to drop (optional; current branch is used " "Name of the branch to drop (optional; current branch is used if omitted)"
"if omitted)"
), ),
) )
branch_drop.add_argument( branch_drop.add_argument(

View File

@@ -20,7 +20,9 @@ def add_mirror_subparsers(subparsers: argparse._SubParsersAction) -> None:
required=True, required=True,
) )
mirror_list = mirror_subparsers.add_parser("list", help="List configured mirrors for repositories") mirror_list = mirror_subparsers.add_parser(
"list", help="List configured mirrors for repositories"
)
add_identifier_arguments(mirror_list) add_identifier_arguments(mirror_list)
mirror_list.add_argument( mirror_list.add_argument(
"--source", "--source",
@@ -29,15 +31,21 @@ def add_mirror_subparsers(subparsers: argparse._SubParsersAction) -> None:
help="Which mirror source to show.", help="Which mirror source to show.",
) )
mirror_diff = mirror_subparsers.add_parser("diff", help="Show differences between config mirrors and MIRRORS file") mirror_diff = mirror_subparsers.add_parser(
"diff", help="Show differences between config mirrors and MIRRORS file"
)
add_identifier_arguments(mirror_diff) add_identifier_arguments(mirror_diff)
mirror_merge = mirror_subparsers.add_parser( mirror_merge = mirror_subparsers.add_parser(
"merge", "merge",
help="Merge mirrors between config and MIRRORS file (example: pkgmgr mirror merge config file --all)", help="Merge mirrors between config and MIRRORS file (example: pkgmgr mirror merge config file --all)",
) )
mirror_merge.add_argument("source", choices=["config", "file"], help="Source of mirrors.") mirror_merge.add_argument(
mirror_merge.add_argument("target", choices=["config", "file"], help="Target of mirrors.") "source", choices=["config", "file"], help="Source of mirrors."
)
mirror_merge.add_argument(
"target", choices=["config", "file"], help="Target of mirrors."
)
add_identifier_arguments(mirror_merge) add_identifier_arguments(mirror_merge)
mirror_merge.add_argument( mirror_merge.add_argument(
"--config-path", "--config-path",

View File

@@ -48,9 +48,6 @@ def add_navigation_subparsers(
"--command", "--command",
nargs=argparse.REMAINDER, nargs=argparse.REMAINDER,
dest="shell_command", dest="shell_command",
help=( help=("The shell command (and its arguments) to execute in each repository"),
"The shell command (and its arguments) to execute in each "
"repository"
),
default=[], default=[],
) )

View File

@@ -53,10 +53,7 @@ def _add_proxy_identifier_arguments(parser: argparse.ArgumentParser) -> None:
parser.add_argument( parser.add_argument(
"identifiers", "identifiers",
nargs="*", nargs="*",
help=( help=("Identifier(s) for repositories. Default: Repository of current folder."),
"Identifier(s) for repositories. "
"Default: Repository of current folder."
),
) )
parser.add_argument( parser.add_argument(
"--all", "--all",
@@ -118,12 +115,7 @@ def _proxy_has_explicit_selection(args: argparse.Namespace) -> bool:
string_filter = getattr(args, "string", "") or "" string_filter = getattr(args, "string", "") or ""
# Proxy commands currently do not support --tag, so it is not checked here. # Proxy commands currently do not support --tag, so it is not checked here.
return bool( return bool(use_all or identifiers or categories or string_filter)
use_all
or identifiers
or categories
or string_filter
)
def _select_repo_for_current_directory( def _select_repo_for_current_directory(
@@ -204,9 +196,7 @@ def maybe_handle_proxy(args: argparse.Namespace, ctx: CLIContext) -> bool:
If the top-level command is one of the proxy subcommands If the top-level command is one of the proxy subcommands
(git / docker / docker compose), handle it here and return True. (git / docker / docker compose), handle it here and return True.
""" """
all_proxy_subcommands = { all_proxy_subcommands = {sub for subs in PROXY_COMMANDS.values() for sub in subs}
sub for subs in PROXY_COMMANDS.values() for sub in subs
}
if args.command not in all_proxy_subcommands: if args.command not in all_proxy_subcommands:
return False return False

View File

@@ -0,0 +1,5 @@
from __future__ import annotations
from .vscode import open_vscode_workspace
__all__ = ["open_vscode_workspace"]

View File

@@ -0,0 +1,34 @@
from __future__ import annotations
from typing import Any, Dict
from pkgmgr.cli.context import CLIContext
from pkgmgr.core.repository.dir import get_repo_dir
Repository = Dict[str, Any]
def resolve_repository_path(repository: Repository, ctx: CLIContext) -> str:
"""
Resolve the filesystem path for a repository.
Priority:
1. Use explicit keys if present (directory / path / workspace / workspace_dir).
2. Fallback to get_repo_dir(...) using the repositories base directory
from the CLI context.
"""
for key in ("directory", "path", "workspace", "workspace_dir"):
value = repository.get(key)
if value:
return value
base_dir = getattr(ctx, "repositories_base_dir", None) or getattr(
ctx, "repositories_dir", None
)
if not base_dir:
raise RuntimeError(
"Cannot resolve repositories base directory from context; "
"expected ctx.repositories_base_dir or ctx.repositories_dir."
)
return get_repo_dir(base_dir, repository)

View File

@@ -0,0 +1,104 @@
from __future__ import annotations
import json
import os
import shutil
from typing import Any, Dict, List
from pkgmgr.cli.context import CLIContext
from pkgmgr.cli.tools.paths import resolve_repository_path
from pkgmgr.core.command.run import run_command
from pkgmgr.core.repository.identifier import get_repo_identifier
Repository = Dict[str, Any]
def _ensure_vscode_cli_available() -> None:
"""
Ensure that the VS Code CLI ('code') is available in PATH.
"""
if shutil.which("code") is None:
raise RuntimeError(
"VS Code CLI ('code') not found in PATH.\n\n"
"Hint:\n"
" Install Visual Studio Code and ensure the 'code' command is available.\n"
" VS Code → Command Palette → 'Shell Command: Install code command in PATH'\n"
)
def _ensure_identifiers_are_filename_safe(identifiers: List[str]) -> None:
"""
Ensure identifiers can be used in a filename.
If an identifier contains '/', it likely means the repository has not yet
been explicitly identified (no short identifier configured).
"""
invalid = [i for i in identifiers if "/" in i or os.sep in i]
if invalid:
raise RuntimeError(
"Cannot create VS Code workspace.\n\n"
"The following repositories are not yet identified "
"(identifier contains '/'): \n"
+ "\n".join(f" - {i}" for i in invalid)
+ "\n\n"
"Hint:\n"
" The repository has no short identifier yet.\n"
" Add an explicit identifier in your configuration before using `pkgmgr tools code`.\n"
)
def _resolve_workspaces_dir(ctx: CLIContext) -> str:
directories_cfg = ctx.config_merged.get("directories") or {}
return os.path.expanduser(directories_cfg.get("workspaces", "~/Workspaces"))
def _build_workspace_filename(identifiers: List[str]) -> str:
sorted_identifiers = sorted(identifiers)
return "_".join(sorted_identifiers) + ".code-workspace"
def _build_workspace_data(
selected: List[Repository], ctx: CLIContext
) -> Dict[str, Any]:
folders = [{"path": resolve_repository_path(repo, ctx)} for repo in selected]
return {
"folders": folders,
"settings": {},
}
def open_vscode_workspace(ctx: CLIContext, selected: List[Repository]) -> None:
"""
Create (if missing) and open a VS Code workspace for the selected repositories.
Policy:
- Fail with a clear error if VS Code CLI is missing.
- Fail with a clear error if any repository identifier contains '/', because that
indicates the repo has not been explicitly identified (no short identifier).
- Do NOT auto-sanitize identifiers and do NOT create subfolders under workspaces.
"""
if not selected:
print("No repositories selected.")
return
_ensure_vscode_cli_available()
identifiers = [get_repo_identifier(repo, ctx.all_repositories) for repo in selected]
_ensure_identifiers_are_filename_safe(identifiers)
workspaces_dir = _resolve_workspaces_dir(ctx)
os.makedirs(workspaces_dir, exist_ok=True)
workspace_name = _build_workspace_filename(identifiers)
workspace_file = os.path.join(workspaces_dir, workspace_name)
workspace_data = _build_workspace_data(selected, ctx)
if not os.path.exists(workspace_file):
with open(workspace_file, "w", encoding="utf-8") as f:
json.dump(workspace_data, f, indent=4)
print(f"Created workspace file: {workspace_file}")
else:
print(f"Using existing workspace file: {workspace_file}")
run_command(f'code "{workspace_file}"')

View File

@@ -2,6 +2,7 @@ import os
import hashlib import hashlib
import re import re
def generate_alias(repo, bin_dir, existing_aliases): def generate_alias(repo, bin_dir, existing_aliases):
""" """
Generate an alias for a repository based on its repository name. Generate an alias for a repository based on its repository name.

View File

@@ -98,8 +98,7 @@ def create_ink(
if alias_name == repo_identifier: if alias_name == repo_identifier:
if not quiet: if not quiet:
print( print(
f"Alias '{alias_name}' equals identifier. " f"Alias '{alias_name}' equals identifier. Skipping alias creation."
"Skipping alias creation."
) )
return return

View File

@@ -8,6 +8,7 @@ class CliLayer(str, Enum):
""" """
CLI layer precedence (lower number = stronger layer). CLI layer precedence (lower number = stronger layer).
""" """
OS_PACKAGES = "os-packages" OS_PACKAGES = "os-packages"
NIX = "nix" NIX = "nix"
PYTHON = "python" PYTHON = "python"

View File

@@ -34,11 +34,7 @@ def _nix_binary_candidates(home: str, names: List[str]) -> List[str]:
""" """
Build possible Nix profile binary paths for a list of candidate names. Build possible Nix profile binary paths for a list of candidate names.
""" """
return [ return [os.path.join(home, ".nix-profile", "bin", name) for name in names if name]
os.path.join(home, ".nix-profile", "bin", name)
for name in names
if name
]
def _path_binary_candidates(names: List[str]) -> List[str]: def _path_binary_candidates(names: List[str]) -> List[str]:
@@ -148,7 +144,8 @@ def resolve_command_for_repo(
# c) Nix profile binaries # c) Nix profile binaries
nix_binaries = [ nix_binaries = [
path for path in _nix_binary_candidates(home, candidate_names) path
for path in _nix_binary_candidates(home, candidate_names)
if _is_executable(path) if _is_executable(path)
] ]
nix_binary = nix_binaries[0] if nix_binaries else None nix_binary = nix_binaries[0] if nix_binaries else None

View File

@@ -51,6 +51,7 @@ Repo = Dict[str, Any]
# Hilfsfunktionen # Hilfsfunktionen
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
def _deep_merge(base: Dict[str, Any], override: Dict[str, Any]) -> Dict[str, Any]: def _deep_merge(base: Dict[str, Any], override: Dict[str, Any]) -> Dict[str, Any]:
""" """
Recursively merge two dictionaries. Recursively merge two dictionaries.
@@ -58,11 +59,7 @@ def _deep_merge(base: Dict[str, Any], override: Dict[str, Any]) -> Dict[str, Any
Values from `override` win over values in `base`. Values from `override` win over values in `base`.
""" """
for key, value in override.items(): for key, value in override.items():
if ( if key in base and isinstance(base[key], dict) and isinstance(value, dict):
key in base
and isinstance(base[key], dict)
and isinstance(value, dict)
):
_deep_merge(base[key], value) _deep_merge(base[key], value)
else: else:
base[key] = value base[key] = value
@@ -93,9 +90,7 @@ def _merge_repo_lists(
- Wenn category_name gesetzt ist, wird dieser in - Wenn category_name gesetzt ist, wird dieser in
repo["category_files"] eingetragen. repo["category_files"] eingetragen.
""" """
index: Dict[Tuple[str, str, str], Repo] = { index: Dict[Tuple[str, str, str], Repo] = {_repo_key(r): r for r in base_list}
_repo_key(r): r for r in base_list
}
for src in new_list: for src in new_list:
key = _repo_key(src) key = _repo_key(src)
@@ -233,10 +228,12 @@ def _load_defaults_from_package_or_project() -> Dict[str, Any]:
return {"directories": {}, "repositories": []} return {"directories": {}, "repositories": []}
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
# Hauptfunktion # Hauptfunktion
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
def load_config(user_config_path: str) -> Dict[str, Any]: def load_config(user_config_path: str) -> Dict[str, Any]:
""" """
Load and merge configuration for pkgmgr. Load and merge configuration for pkgmgr.
@@ -289,8 +286,12 @@ def load_config(user_config_path: str) -> Dict[str, Any]:
# repositories # repositories
merged["repositories"] = [] merged["repositories"] = []
_merge_repo_lists(merged["repositories"], defaults["repositories"], category_name=None) _merge_repo_lists(
_merge_repo_lists(merged["repositories"], user_cfg["repositories"], category_name=None) merged["repositories"], defaults["repositories"], category_name=None
)
_merge_repo_lists(
merged["repositories"], user_cfg["repositories"], category_name=None
)
# andere Top-Level-Keys (falls vorhanden) # andere Top-Level-Keys (falls vorhanden)
other_keys = (set(defaults.keys()) | set(user_cfg.keys())) - { other_keys = (set(defaults.keys()) | set(user_cfg.keys())) - {

View File

@@ -1,9 +1,10 @@
import yaml import yaml
import os import os
def save_user_config(user_config, USER_CONFIG_PATH: str): def save_user_config(user_config, USER_CONFIG_PATH: str):
"""Save the user configuration to USER_CONFIG_PATH.""" """Save the user configuration to USER_CONFIG_PATH."""
os.makedirs(os.path.dirname(USER_CONFIG_PATH), exist_ok=True) os.makedirs(os.path.dirname(USER_CONFIG_PATH), exist_ok=True)
with open(USER_CONFIG_PATH, 'w') as f: with open(USER_CONFIG_PATH, "w") as f:
yaml.dump(user_config, f) yaml.dump(user_config, f)
print(f"User configuration updated in {USER_CONFIG_PATH}.") print(f"User configuration updated in {USER_CONFIG_PATH}.")

View File

@@ -16,7 +16,9 @@ class EnvTokenProvider:
source_name: str = "env" source_name: str = "env"
def get(self, request: TokenRequest) -> Optional[TokenResult]: def get(self, request: TokenRequest) -> Optional[TokenResult]:
for key in env_var_candidates(request.provider_kind, request.host, request.owner): for key in env_var_candidates(
request.provider_kind, request.host, request.owner
):
val = os.environ.get(key) val = os.environ.get(key)
if val: if val:
return TokenResult(token=val.strip(), source=self.source_name) return TokenResult(token=val.strip(), source=self.source_name)

View File

@@ -15,6 +15,7 @@ class GhTokenProvider:
This does NOT persist anything; it only reads what `gh` already knows. This does NOT persist anything; it only reads what `gh` already knows.
""" """
source_name: str = "gh" source_name: str = "gh"
def get(self, request: TokenRequest) -> Optional[TokenResult]: def get(self, request: TokenRequest) -> Optional[TokenResult]:

View File

@@ -21,9 +21,7 @@ def _import_keyring():
try: try:
import keyring # type: ignore import keyring # type: ignore
except Exception as exc: # noqa: BLE001 except Exception as exc: # noqa: BLE001
raise KeyringUnavailableError( raise KeyringUnavailableError("python-keyring is not installed.") from exc
"python-keyring is not installed."
) from exc
# Some environments have keyring installed but no usable backend. # Some environments have keyring installed but no usable backend.
# We do a lightweight "backend sanity check" by attempting to read the backend. # We do a lightweight "backend sanity check" by attempting to read the backend.

View File

@@ -9,7 +9,12 @@ from .providers.env import EnvTokenProvider
from .providers.gh import GhTokenProvider from .providers.gh import GhTokenProvider
from .providers.keyring import KeyringTokenProvider from .providers.keyring import KeyringTokenProvider
from .providers.prompt import PromptTokenProvider from .providers.prompt import PromptTokenProvider
from .types import KeyringUnavailableError, NoCredentialsError, TokenRequest, TokenResult from .types import (
KeyringUnavailableError,
NoCredentialsError,
TokenRequest,
TokenResult,
)
from .validate import validate_token from .validate import validate_token
@@ -55,7 +60,10 @@ class TokenResolver:
print(f" {msg}", file=sys.stderr) print(f" {msg}", file=sys.stderr)
print(" Tokens will NOT be persisted securely.", file=sys.stderr) print(" Tokens will NOT be persisted securely.", file=sys.stderr)
print("", file=sys.stderr) print("", file=sys.stderr)
print(" To enable secure token storage, install python-keyring:", file=sys.stderr) print(
" To enable secure token storage, install python-keyring:",
file=sys.stderr,
)
print(" pip install keyring", file=sys.stderr) print(" pip install keyring", file=sys.stderr)
print("", file=sys.stderr) print("", file=sys.stderr)
print(" Or install via system packages:", file=sys.stderr) print(" Or install via system packages:", file=sys.stderr)

View File

@@ -13,7 +13,9 @@ class KeyringKey:
username: str username: str
def build_keyring_key(provider_kind: str, host: str, owner: Optional[str]) -> KeyringKey: def build_keyring_key(
provider_kind: str, host: str, owner: Optional[str]
) -> KeyringKey:
"""Build a stable keyring key. """Build a stable keyring key.
- service: "pkgmgr:<provider>" - service: "pkgmgr:<provider>"
@@ -21,11 +23,15 @@ def build_keyring_key(provider_kind: str, host: str, owner: Optional[str]) -> Ke
""" """
provider_kind = str(provider_kind).strip().lower() provider_kind = str(provider_kind).strip().lower()
host = str(host).strip() host = str(host).strip()
owner_part = (str(owner).strip() if owner else "-") owner_part = str(owner).strip() if owner else "-"
return KeyringKey(service=f"pkgmgr:{provider_kind}", username=f"{host}|{owner_part}") return KeyringKey(
service=f"pkgmgr:{provider_kind}", username=f"{host}|{owner_part}"
)
def env_var_candidates(provider_kind: str, host: str, owner: Optional[str]) -> list[str]: def env_var_candidates(
provider_kind: str, host: str, owner: Optional[str]
) -> list[str]:
"""Return a list of environment variable names to try. """Return a list of environment variable names to try.
Order is from most specific to most generic. Order is from most specific to most generic.

View File

@@ -1,6 +1,6 @@
from __future__ import annotations from __future__ import annotations
from .errors import GitError from .errors import GitRunError
from .run import run from .run import run
""" """
@@ -12,6 +12,6 @@ details of subprocess handling.
""" """
__all__ = [ __all__ = [
"GitError", "GitRunError",
"run", "run",
] ]

View File

@@ -16,7 +16,7 @@ from .fetch import GitFetchError, fetch
from .init import GitInitError, init from .init import GitInitError, init
from .merge_no_ff import GitMergeError, merge_no_ff from .merge_no_ff import GitMergeError, merge_no_ff
from .pull import GitPullError, pull from .pull import GitPullError, pull
from .pull_args import GitPullArgsError, pull_args # <-- add from .pull_args import GitPullArgsError, pull_args
from .pull_ff_only import GitPullFfOnlyError, pull_ff_only from .pull_ff_only import GitPullFfOnlyError, pull_ff_only
from .push import GitPushError, push from .push import GitPushError, push
from .push_upstream import GitPushUpstreamError, push_upstream from .push_upstream import GitPushUpstreamError, push_upstream
@@ -30,7 +30,7 @@ __all__ = [
"fetch", "fetch",
"checkout", "checkout",
"pull", "pull",
"pull_args", # <-- add "pull_args",
"pull_ff_only", "pull_ff_only",
"merge_no_ff", "merge_no_ff",
"push", "push",
@@ -52,7 +52,7 @@ __all__ = [
"GitFetchError", "GitFetchError",
"GitCheckoutError", "GitCheckoutError",
"GitPullError", "GitPullError",
"GitPullArgsError", # <-- add "GitPullArgsError",
"GitPullFfOnlyError", "GitPullFfOnlyError",
"GitMergeError", "GitMergeError",
"GitPushError", "GitPushError",

View File

@@ -2,7 +2,7 @@ from __future__ import annotations
from typing import Iterable, List, Sequence, Union from typing import Iterable, List, Sequence, Union
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -37,7 +37,7 @@ def add(
try: try:
run(["add", *normalized], cwd=cwd, preview=preview) run(["add", *normalized], cwd=cwd, preview=preview)
except GitError as exc: except GitRunError as exc:
raise GitAddError( raise GitAddError(
f"Failed to add paths to staging area: {normalized!r}.", f"Failed to add paths to staging area: {normalized!r}.",
cwd=cwd, cwd=cwd,

View File

@@ -1,7 +1,6 @@
# src/pkgmgr/core/git/commands/add_all.py
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitCommandError, GitRunError
from ..run import run from ..run import run
@@ -18,5 +17,7 @@ def add_all(*, cwd: str = ".", preview: bool = False) -> None:
""" """
try: try:
run(["add", "-A"], cwd=cwd, preview=preview) run(["add", "-A"], cwd=cwd, preview=preview)
except GitError as exc: except GitRunError as exc:
raise GitAddAllError("Failed to stage all changes with `git add -A`.", cwd=cwd) from exc raise GitAddAllError(
"Failed to stage all changes with `git add -A`.", cwd=cwd
) from exc

View File

@@ -1,6 +1,6 @@
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -27,7 +27,7 @@ def add_remote(
cwd=cwd, cwd=cwd,
preview=preview, preview=preview,
) )
except GitError as exc: except GitRunError as exc:
raise GitAddRemoteError( raise GitAddRemoteError(
f"Failed to add remote {name!r} with URL {url!r}.", f"Failed to add remote {name!r} with URL {url!r}.",
cwd=cwd, cwd=cwd,

View File

@@ -1,6 +1,6 @@
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -27,7 +27,7 @@ def add_remote_push_url(
cwd=cwd, cwd=cwd,
preview=preview, preview=preview,
) )
except GitError as exc: except GitRunError as exc:
raise GitAddRemotePushUrlError( raise GitAddRemotePushUrlError(
f"Failed to add push url {url!r} to remote {remote!r}.", f"Failed to add push url {url!r} to remote {remote!r}.",
cwd=cwd, cwd=cwd,

View File

@@ -1,7 +1,6 @@
# src/pkgmgr/core/git/commands/branch_move.py
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -18,5 +17,7 @@ def branch_move(branch: str, *, cwd: str = ".", preview: bool = False) -> None:
""" """
try: try:
run(["branch", "-M", branch], cwd=cwd, preview=preview) run(["branch", "-M", branch], cwd=cwd, preview=preview)
except GitError as exc: except GitRunError as exc:
raise GitBranchMoveError(f"Failed to move/rename current branch to {branch!r}.", cwd=cwd) from exc raise GitBranchMoveError(
f"Failed to move/rename current branch to {branch!r}.", cwd=cwd
) from exc

View File

@@ -1,6 +1,6 @@
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -11,7 +11,7 @@ class GitCheckoutError(GitCommandError):
def checkout(branch: str, cwd: str = ".") -> None: def checkout(branch: str, cwd: str = ".") -> None:
try: try:
run(["checkout", branch], cwd=cwd) run(["checkout", branch], cwd=cwd)
except GitError as exc: except GitRunError as exc:
raise GitCheckoutError( raise GitCheckoutError(
f"Failed to checkout branch {branch!r}.", f"Failed to checkout branch {branch!r}.",
cwd=cwd, cwd=cwd,

View File

@@ -2,7 +2,7 @@ from __future__ import annotations
from typing import List from typing import List
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -25,7 +25,7 @@ def clone(
""" """
try: try:
run(["clone", *args], cwd=cwd, preview=preview) run(["clone", *args], cwd=cwd, preview=preview)
except GitError as exc: except GitRunError as exc:
raise GitCloneError( raise GitCloneError(
f"Git clone failed with args={args!r}.", f"Git clone failed with args={args!r}.",
cwd=cwd, cwd=cwd,

View File

@@ -1,6 +1,6 @@
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -30,7 +30,7 @@ def commit(
try: try:
run(args, cwd=cwd, preview=preview) run(args, cwd=cwd, preview=preview)
except GitError as exc: except GitRunError as exc:
raise GitCommitError( raise GitCommitError(
"Failed to create commit.", "Failed to create commit.",
cwd=cwd, cwd=cwd,

View File

@@ -1,6 +1,6 @@
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -16,7 +16,7 @@ def create_branch(branch: str, base: str, cwd: str = ".") -> None:
""" """
try: try:
run(["checkout", "-b", branch, base], cwd=cwd) run(["checkout", "-b", branch, base], cwd=cwd)
except GitError as exc: except GitRunError as exc:
raise GitCreateBranchError( raise GitCreateBranchError(
f"Failed to create branch {branch!r} from base {base!r}.", f"Failed to create branch {branch!r} from base {base!r}.",
cwd=cwd, cwd=cwd,

View File

@@ -1,6 +1,6 @@
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -12,7 +12,7 @@ def delete_local_branch(branch: str, cwd: str = ".", force: bool = False) -> Non
flag = "-D" if force else "-d" flag = "-D" if force else "-d"
try: try:
run(["branch", flag, branch], cwd=cwd) run(["branch", flag, branch], cwd=cwd)
except GitError as exc: except GitRunError as exc:
raise GitDeleteLocalBranchError( raise GitDeleteLocalBranchError(
f"Failed to delete local branch {branch!r} (flag {flag}).", f"Failed to delete local branch {branch!r} (flag {flag}).",
cwd=cwd, cwd=cwd,

View File

@@ -1,6 +1,6 @@
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -11,7 +11,7 @@ class GitDeleteRemoteBranchError(GitCommandError):
def delete_remote_branch(remote: str, branch: str, cwd: str = ".") -> None: def delete_remote_branch(remote: str, branch: str, cwd: str = ".") -> None:
try: try:
run(["push", remote, "--delete", branch], cwd=cwd) run(["push", remote, "--delete", branch], cwd=cwd)
except GitError as exc: except GitRunError as exc:
raise GitDeleteRemoteBranchError( raise GitDeleteRemoteBranchError(
f"Failed to delete remote branch {branch!r} on {remote!r}.", f"Failed to delete remote branch {branch!r} on {remote!r}.",
cwd=cwd, cwd=cwd,

View File

@@ -1,6 +1,6 @@
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -33,7 +33,7 @@ def fetch(
try: try:
run(args, cwd=cwd, preview=preview) run(args, cwd=cwd, preview=preview)
except GitError as exc: except GitRunError as exc:
raise GitFetchError( raise GitFetchError(
f"Failed to fetch from remote {remote!r}.", f"Failed to fetch from remote {remote!r}.",
cwd=cwd, cwd=cwd,

View File

@@ -1,7 +1,6 @@
# src/pkgmgr/core/git/commands/init.py
from __future__ import annotations from __future__ import annotations
from ..errors import GitError, GitCommandError from ..errors import GitRunError, GitCommandError
from ..run import run from ..run import run
@@ -18,5 +17,5 @@ def init(*, cwd: str = ".", preview: bool = False) -> None:
""" """
try: try:
run(["init"], cwd=cwd, preview=preview) run(["init"], cwd=cwd, preview=preview)
except GitError as exc: except GitRunError as exc:
raise GitInitError("Failed to initialize git repository.", cwd=cwd) from exc raise GitInitError("Failed to initialize git repository.", cwd=cwd) from exc

Some files were not shown because too many files have changed in this diff Show More