Package pkgmgr with multi-format manifests

- Add pyproject.toml and setuptools config for pkgmgr packaging
- Add Nix flake (devShell + pkgmgr package output)
- Add Arch PKGBUILD for system packaging
- Introduce pkgmgr.yml manifest for repo-level dependencies
- Refactor CLI into pkgmgr/cli.py and make main.py a thin entrypoint
- Extend install_repos to handle pkgmgr.yml, PKGBUILD, flake.nix, Ansible and Python manifests
- Enhance status/update to show Nix/yay system status and upgrades
- Improve .gitignore and document requirements.yml

Created with AI (ChatGPT) – see conversation: https://chatgpt.com/share/6932f2ca-f560-800f-8bb0-52cb82f27e88
This commit is contained in:
Kevin Veen-Birkenbach
2025-12-05 15:57:45 +01:00
parent c4395a4764
commit 41084234c7
13 changed files with 956 additions and 546 deletions

27
.gitignore vendored
View File

@@ -1,2 +1,27 @@
# Prevents unwanted files from being committed to version control.
# Custom Config file
config/config.yaml
*__pycache__*
# Python bytecode
__pycache__/
*.pyc
# Virtual environments
.venv/
venv/
.venvs/
# Build artifacts
dist/
build/
*.egg-info/
# Editor files
.vscode/
.idea/
# OS noise
.DS_Store
Thumbs.db

39
PKGBUILD Normal file
View File

@@ -0,0 +1,39 @@
# Maintainer: Kevin Veen-Birkenbach <info@veen.world>
pkgname=package-manager
pkgver=0.1.0
pkgrel=1
pkgdesc="A configurable Python tool to manage multiple repositories via Bash and automate common Git operations."
arch=('any')
url="https://github.com/kevinveenbirkenbach/package-manager"
license=('MIT')
depends=(
'python'
'python-yaml'
'git'
'bash'
)
makedepends=(
'python-build'
'python-installer'
'python-wheel'
'python-setuptools'
)
source=("$pkgname-$pkgver.tar.gz::$url/archive/refs/tags/v$pkgver.tar.gz")
sha256sums=('SKIP')
build() {
cd "$srcdir/$pkgname-$pkgver"
python -m build --wheel --no-isolation
}
package() {
cd "$srcdir/$pkgname-$pkgver"
python -m installer --destdir="$pkgdir" dist/*.whl
# Optional: add pkgmgr executable symlink
install -Dm755 main.py "$pkgdir/usr/bin/pkgmgr"
}

0
__init__.py Normal file
View File

40
flake.nix Normal file
View File

@@ -0,0 +1,40 @@
# flake.nix
# This file defines a Nix flake providing a reproducible development environment
# and optional installation package for the package-manager tool.
{
description = "Nix flake for Kevin's package-manager tool";
inputs = {
nixpkgs.url = "github:NixOS/nixpkgs/nixos-unstable";
};
outputs = { self, nixpkgs }:
let
pkgs = nixpkgs.legacyPackages.x86_64-linux;
in {
# Development environment used via: nix develop
devShells.default = pkgs.mkShell {
# System packages for development
buildInputs = [
pkgs.python311
pkgs.python311Packages.pyyaml
pkgs.git
];
# Message shown on environment entry
shellHook = ''
echo "Entered pkgmgr development environment";
'';
};
# Optional installable package for "nix profile install"
packages.pkgmgr = pkgs.python311Packages.buildPythonApplication {
pname = "package-manager";
version = "0.1.0";
src = ./.;
propagatedBuildInputs = [ pkgs.python311Packages.pyyaml ];
};
};
}

458
main.py
View File

@@ -1,460 +1,6 @@
#!/usr/bin/env python3
import os
import yaml
import argparse
import json
import os
import sys
from pkgmgr.cli import main
# Define configuration file paths.
USER_CONFIG_PATH = os.path.join(os.path.dirname(os.path.realpath(__file__)), "config", "config.yaml")
from pkgmgr.clone_repos import clone_repos
from pkgmgr.config_init import config_init
from pkgmgr.create_ink import create_ink
from pkgmgr.deinstall_repos import deinstall_repos
from pkgmgr.delete_repos import delete_repos
from pkgmgr.exec_proxy_command import exec_proxy_command
from pkgmgr.filter_ignored import filter_ignored
from pkgmgr.get_repo_identifier import get_repo_identifier
from pkgmgr.get_selected_repos import get_selected_repos
from pkgmgr.install_repos import install_repos
from pkgmgr.interactive_add import interactive_add
from pkgmgr.list_repositories import list_repositories
from pkgmgr.load_config import load_config
from pkgmgr.resolve_repos import resolve_repos
from pkgmgr.run_command import run_command
from pkgmgr.save_user_config import save_user_config
from pkgmgr.show_config import show_config
from pkgmgr.status_repos import status_repos
from pkgmgr.update_repos import update_repos
# Commands proxied by package-manager
PROXY_COMMANDS = {
"git":[
"pull",
"push",
"diff",
"add",
"show",
"checkout",
"clone",
"reset",
"revert",
"rebase",
"commit"
],
"docker":[
"start",
"stop",
"build"
],
"docker compose":[
"up",
"down",
"exec",
"ps",
"restart",
]
}
class SortedSubParsersAction(argparse._SubParsersAction):
def add_parser(self, name, **kwargs):
parser = super().add_parser(name, **kwargs)
# Sort the list of subparsers each time one is added
self._choices_actions.sort(key=lambda a: a.dest)
return parser
# Main program.
if __name__ == "__main__":
CONFIG_MERGED = load_config(USER_CONFIG_PATH)
REPOSITORIES_BASE_DIR = os.path.expanduser(CONFIG_MERGED["directories"]["repositories"])
ALL_REPOSITORIES = CONFIG_MERGED["repositories"]
BINARIES_DIRECTORY = os.path.expanduser(CONFIG_MERGED["directories"]["binaries"])
description_text = """\
\033[1;32mPackage Manager 🤖📦\033[0m
\033[3mKevin's Package Manager ist drafted by and designed for:
\033[1;34mKevin Veen-Birkenbach
\033[0m\033[4mhttps://www.veen.world/\033[0m
\033[1mOverview:\033[0m
A configurable Python tool to manage multiple repositories via a unified command-line interface.
This tool automates common Git operations (clone, pull, push, status, etc.) and creates executable wrappers and custom aliases to simplify your workflow.
\033[1mFeatures:\033[0m
\033[1;33mAuto-install & Setup:\033[0m Automatically detect and set up repositories.
\033[1;33mGit Command Integration:\033[0m Execute Git commands with extra parameters.
\033[1;33mExplorer & Terminal Support:\033[0m Open repositories in your file manager or a new terminal tab.
\033[1;33mComprehensive Configuration:\033[0m Manage settings via YAML files (default & user-specific).
For detailed help on each command, use:
\033[1m pkgmgr <command> --help\033[0m
"""
parser = argparse.ArgumentParser(description=description_text,formatter_class=argparse.RawTextHelpFormatter)
subparsers = parser.add_subparsers(dest="command", help="Subcommands", action=SortedSubParsersAction)
def add_identifier_arguments(subparser):
subparser.add_argument(
"identifiers",
nargs="*",
help="Identifier(s) for repositories. Default: Repository of current folder.",
)
subparser.add_argument(
"--all",
action="store_true",
default=False,
help="Apply the subcommand to all repositories in the config. Some subcommands ask for confirmation. If you want to give this confirmation for all repositories, pipe 'yes'. E.g: yes | pkgmgr {subcommand} --all"
)
subparser.add_argument("--preview", action="store_true", help="Preview changes without executing commands")
subparser.add_argument("--list", action="store_true", help="List affected repositories (with preview or status)")
subparser.add_argument("-a", "--args", nargs=argparse.REMAINDER, dest="extra_args", help="Additional parameters to be attached.",default=[])
def add_install_update_arguments(subparser):
add_identifier_arguments(subparser)
subparser.add_argument(
"-q",
"--quiet",
action="store_true",
help="Suppress warnings and info messages",
)
subparser.add_argument(
"--no-verification",
action="store_true",
default=False,
help="Disable verification via commit/gpg",
)
subparser.add_argument(
"--dependencies",
action="store_true",
help="Also pull and update dependencies",
)
subparser.add_argument(
"--clone-mode",
choices=["ssh", "https", "shallow"],
default="ssh",
help="Specify the clone mode: ssh, https, or shallow (HTTPS shallow clone; default: ssh)",
)
install_parser = subparsers.add_parser("install", help="Setup repository/repositories alias links to executables")
add_install_update_arguments(install_parser)
update_parser = subparsers.add_parser("update", help="Update (pull + install) repository/repositories")
add_install_update_arguments(update_parser)
update_parser.add_argument("--system", action="store_true", help="Include system update commands")
deinstall_parser = subparsers.add_parser("deinstall", help="Remove alias links to repository/repositories")
add_identifier_arguments(deinstall_parser)
delete_parser = subparsers.add_parser("delete", help="Delete repository/repositories alias links to executables")
add_identifier_arguments(delete_parser)
# Add the 'create' subcommand (with existing identifier arguments)
create_parser = subparsers.add_parser(
"create",
help="Create new repository entries: add them to the config if not already present, initialize the local repository, and push remotely if --remote is set."
)
# Reuse the common identifier arguments
add_identifier_arguments(create_parser)
create_parser.add_argument(
"--remote",
action="store_true",
help="If set, add the remote and push the initial commit."
)
status_parser = subparsers.add_parser("status", help="Show status for repository/repositories or system")
add_identifier_arguments(status_parser)
status_parser.add_argument("--system", action="store_true", help="Show system status")
config_parser = subparsers.add_parser("config", help="Manage configuration")
config_subparsers = config_parser.add_subparsers(dest="subcommand", help="Config subcommands", required=True)
config_show = config_subparsers.add_parser("show", help="Show configuration")
add_identifier_arguments(config_show)
config_add = config_subparsers.add_parser("add", help="Interactively add a new repository entry")
config_edit = config_subparsers.add_parser("edit", help="Edit configuration file with nano")
config_init_parser = config_subparsers.add_parser("init", help="Initialize user configuration by scanning the base directory")
config_delete = config_subparsers.add_parser("delete", help="Delete repository entry from user config")
add_identifier_arguments(config_delete)
config_ignore = config_subparsers.add_parser("ignore", help="Set ignore flag for repository entries in user config")
add_identifier_arguments(config_ignore)
config_ignore.add_argument("--set", choices=["true", "false"], required=True, help="Set ignore to true or false")
path_parser = subparsers.add_parser("path", help="Print the path(s) of repository/repositories")
add_identifier_arguments(path_parser)
explore_parser = subparsers.add_parser("explore", help="Open repository in Nautilus file manager")
add_identifier_arguments(explore_parser)
terminal_parser = subparsers.add_parser("terminal", help="Open repository in a new GNOME Terminal tab")
add_identifier_arguments(terminal_parser)
release_parser = subparsers.add_parser(
"release",
help="Create a release for repository/ies by incrementing version and updating the changelog."
)
release_parser.add_argument(
"release_type",
choices=["major", "minor", "patch"],
help="Type of version increment for the release (major, minor, patch)."
)
release_parser.add_argument(
"-m", "--message",
default="",
help="Optional release message to add to the changelog and tag."
)
add_identifier_arguments(release_parser)
code_parser = subparsers.add_parser("code", help="Open repository workspace with VS Code")
add_identifier_arguments(code_parser)
list_parser = subparsers.add_parser("list", help="List all repositories with details and status")
list_parser.add_argument("--search", default="", help="Filter repositories that contain the given string")
list_parser.add_argument("--status", type=str, default="", help="Filter repositories by status (case insensitive)")
# Add the subcommand parser for "shell"
shell_parser = subparsers.add_parser("shell", help="Execute a shell command in each repository")
add_identifier_arguments(shell_parser)
shell_parser.add_argument("-c", "--command", nargs=argparse.REMAINDER, dest="shell_command", help="The shell command (and its arguments) to execute in each repository",default=[])
make_parser = subparsers.add_parser("make", help="Executes make commands")
add_identifier_arguments(make_parser)
make_subparsers = make_parser.add_subparsers(dest="subcommand", help="Make subcommands", required=True)
make_install = make_subparsers.add_parser("install", help="Executes the make install command")
add_identifier_arguments(make_install)
make_deinstall = make_subparsers.add_parser("deinstall", help="Executes the make deinstall command")
proxy_command_parsers = {}
for command, subcommands in PROXY_COMMANDS.items():
for subcommand in subcommands:
proxy_command_parsers[f"{command}_{subcommand}"] = subparsers.add_parser(
subcommand,
help=f"Proxies '{command} {subcommand}' to repository/ies",
description=f"Executes '{command} {subcommand}' for the identified repos.\nTo recieve more help execute '{command} {subcommand} --help'",
formatter_class=argparse.RawTextHelpFormatter
)
if subcommand in ["pull", "clone"]:
proxy_command_parsers[f"{command}_{subcommand}"].add_argument(
"--no-verification",
action="store_true",
default=False,
help="Disable verification via commit/gpg",
)
if subcommand == "clone":
proxy_command_parsers[f"{command}_{subcommand}"].add_argument(
"--clone-mode",
choices=["ssh", "https", "shallow"],
default="ssh",
help="Specify the clone mode: ssh, https, or shallow (HTTPS shallow clone; default: ssh)",
)
add_identifier_arguments(proxy_command_parsers[f"{command}_{subcommand}"])
args = parser.parse_args()
# All
if args.command and not args.command in ["config","list","create"]:
selected = get_selected_repos(args.all,ALL_REPOSITORIES,args.identifiers)
for command, subcommands in PROXY_COMMANDS.items():
for subcommand in subcommands:
if args.command == subcommand:
if args.command == "clone":
clone_repos(
selected,
REPOSITORIES_BASE_DIR,
ALL_REPOSITORIES,
args.preview,
args.no_verification,
args.clone_mode
)
elif args.command == "pull":
from pkgmgr.pull_with_verification import pull_with_verification
pull_with_verification(
selected,
REPOSITORIES_BASE_DIR,
ALL_REPOSITORIES,
args.extra_args,
args.no_verification,
args.preview
)
else:
exec_proxy_command(command,selected, REPOSITORIES_BASE_DIR, ALL_REPOSITORIES, args.command, args.extra_args, args.preview)
exit(0)
if args.command in ["make"]:
exec_proxy_command(args.command,selected, REPOSITORIES_BASE_DIR, ALL_REPOSITORIES, args.subcommand, args.extra_args, args.preview)
exit(0)
# Dispatch commands.
if args.command == "install":
install_repos(
selected,
REPOSITORIES_BASE_DIR,
BINARIES_DIRECTORY,
ALL_REPOSITORIES,
args.no_verification,
args.preview,
args.quiet,
args.clone_mode,
args.dependencies,
)
elif args.command == "create":
from pkgmgr.create_repo import create_repo
# If no identifiers are provided, you can decide to either use the repository of the current folder
# or prompt the user to supply at least one identifier.
if not args.identifiers:
print("No identifiers provided. Please specify at least one identifier in the format provider/account/repository.")
sys.exit(1)
else:
selected = get_selected_repos(True,ALL_REPOSITORIES,None)
for identifier in args.identifiers:
create_repo(identifier, CONFIG_MERGED, USER_CONFIG_PATH, BINARIES_DIRECTORY, remote=args.remote, preview=args.preview)
elif args.command == "list":
list_repositories(ALL_REPOSITORIES, REPOSITORIES_BASE_DIR, BINARIES_DIRECTORY, search_filter=args.search, status_filter=args.status)
elif args.command == "deinstall":
deinstall_repos(selected,REPOSITORIES_BASE_DIR, BINARIES_DIRECTORY, ALL_REPOSITORIES, preview=args.preview)
elif args.command == "delete":
delete_repos(selected,REPOSITORIES_BASE_DIR, ALL_REPOSITORIES, preview=args.preview)
elif args.command == "update":
update_repos(
selected,
REPOSITORIES_BASE_DIR,
BINARIES_DIRECTORY,
ALL_REPOSITORIES,
args.no_verification,
args.system,
args.preview,
args.quiet,
args.dependencies,
args.clone_mode
)
elif args.command == "release":
if not selected:
print("No repositories selected for release.")
exit(1)
# Import the release function from pkgmgr/release.py
from pkgmgr import release as rel
# Save the original working directory.
original_dir = os.getcwd()
for repo in selected:
# Determine the repository directory
repo_dir = repo.get("directory")
if not repo_dir:
from pkgmgr.get_repo_dir import get_repo_dir
repo_dir = get_repo_dir(REPOSITORIES_BASE_DIR, repo)
# Dynamically determine the file paths for pyproject.toml and CHANGELOG.md.
pyproject_path = os.path.join(repo_dir, "pyproject.toml")
changelog_path = os.path.join(repo_dir, "CHANGELOG.md")
print(f"Releasing repository '{repo.get('repository')}' in '{repo_dir}'...")
# Change into the repository directory so Git commands run in the right context.
os.chdir(repo_dir)
# Call the release function with the proper parameters.
rel.release(
pyproject_path=pyproject_path,
changelog_path=changelog_path,
release_type=args.release_type,
message=args.message
)
# Change back to the original working directory.
os.chdir(original_dir)
elif args.command == "status":
status_repos(selected,REPOSITORIES_BASE_DIR, ALL_REPOSITORIES, args.extra_args, list_only=args.list, system_status=args.system, preview=args.preview)
elif args.command == "explore":
for repository in selected:
run_command(f"nautilus {repository['directory']} & disown")
elif args.command == "code":
if not selected:
print("No repositories selected.")
else:
identifiers = [get_repo_identifier(repo, ALL_REPOSITORIES) for repo in selected]
sorted_identifiers = sorted(identifiers)
workspace_name = "_".join(sorted_identifiers) + ".code-workspace"
workspaces_dir = os.path.expanduser(CONFIG_MERGED.get("directories").get("workspaces"))
os.makedirs(workspaces_dir, exist_ok=True)
workspace_file = os.path.join(workspaces_dir, workspace_name)
folders = []
for repository in selected:
folders.append({"path": repository["directory"]})
workspace_data = {
"folders": folders,
"settings": {}
}
if not os.path.exists(workspace_file):
with open(workspace_file, "w") as f:
json.dump(workspace_data, f, indent=4)
print(f"Created workspace file: {workspace_file}")
else:
print(f"Using existing workspace file: {workspace_file}")
run_command(f'code "{workspace_file}"')
elif args.command == "terminal":
for repository in selected:
run_command(f'gnome-terminal --tab --working-directory="{repository["directory"]}"')
elif args.command == "path":
for repository in selected:
print(repository["directory"])
elif args.command == "shell":
if not args.shell_command:
print("No shell command specified.")
exit(2)
# Join the provided shell command parts into one string.
command_to_run = " ".join(args.shell_command)
for repository in selected:
print(f"Executing in '{repository['directory']}': {command_to_run}")
run_command(command_to_run, cwd=repository["directory"], preview=args.preview)
elif args.command == "config":
if args.subcommand == "show":
if args.all or (not args.identifiers):
show_config([], USER_CONFIG_PATH, full_config=True)
else:
selected = resolve_repos(args.identifiers, ALL_REPOSITORIES)
if selected:
show_config(selected, USER_CONFIG_PATH, full_config=False)
elif args.subcommand == "add":
interactive_add(CONFIG_MERGED,USER_CONFIG_PATH)
elif args.subcommand == "edit":
"""Open the user configuration file in nano."""
run_command(f"nano {USER_CONFIG_PATH}")
elif args.subcommand == "init":
if os.path.exists(USER_CONFIG_PATH):
with open(USER_CONFIG_PATH, 'r') as f:
user_config = yaml.safe_load(f) or {}
else:
user_config = {"repositories": []}
config_init(user_config, CONFIG_MERGED, BINARIES_DIRECTORY, USER_CONFIG_PATH)
elif args.subcommand == "delete":
# Load user config from USER_CONFIG_PATH.
if os.path.exists(USER_CONFIG_PATH):
with open(USER_CONFIG_PATH, 'r') as f:
user_config = yaml.safe_load(f) or {"repositories": []}
else:
user_config = {"repositories": []}
if args.all or not args.identifiers:
print("You must specify identifiers to delete.")
else:
to_delete = resolve_repos(args.identifiers, user_config.get("repositories", []))
new_repos = [entry for entry in user_config.get("repositories", []) if entry not in to_delete]
user_config["repositories"] = new_repos
save_user_config(user_config,USER_CONFIG_PATH)
print(f"Deleted {len(to_delete)} entries from user config.")
elif args.subcommand == "ignore":
# Load user config from USER_CONFIG_PATH.
if os.path.exists(USER_CONFIG_PATH):
with open(USER_CONFIG_PATH, 'r') as f:
user_config = yaml.safe_load(f) or {"repositories": []}
else:
user_config = {"repositories": []}
if args.all or not args.identifiers:
print("You must specify identifiers to modify ignore flag.")
else:
to_modify = resolve_repos(args.identifiers, user_config.get("repositories", []))
for entry in user_config["repositories"]:
key = (entry.get("provider"), entry.get("account"), entry.get("repository"))
for mod in to_modify:
mod_key = (mod.get("provider"), mod.get("account"), mod.get("repository"))
if key == mod_key:
entry["ignore"] = (args.set == "true")
print(f"Set ignore for {key} to {entry['ignore']}")
save_user_config(user_config,USER_CONFIG_PATH)
else:
parser.print_help()
main()

7
pkgmgr.yml Normal file
View File

@@ -0,0 +1,7 @@
version: 1
author: "Kevin Veen-Birkenbach"
url: "https://github.com/kevinveenbirkenbach/package-manager"
description: "A configurable Python-based package manager for managing multiple repositories via Bash."
dependencies: []

489
pkgmgr/cli.py Executable file
View File

@@ -0,0 +1,489 @@
#!/usr/bin/env python3
import os
import yaml
import argparse
import json
import os
import sys
# Define configuration file paths.
PROJECT_ROOT = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
USER_CONFIG_PATH = os.path.join(PROJECT_ROOT, "config", "config.yaml")
from pkgmgr.clone_repos import clone_repos
from pkgmgr.config_init import config_init
from pkgmgr.create_ink import create_ink
from pkgmgr.deinstall_repos import deinstall_repos
from pkgmgr.delete_repos import delete_repos
from pkgmgr.exec_proxy_command import exec_proxy_command
from pkgmgr.filter_ignored import filter_ignored
from pkgmgr.get_repo_identifier import get_repo_identifier
from pkgmgr.get_selected_repos import get_selected_repos
from pkgmgr.install_repos import install_repos
from pkgmgr.interactive_add import interactive_add
from pkgmgr.list_repositories import list_repositories
from pkgmgr.load_config import load_config
from pkgmgr.resolve_repos import resolve_repos
from pkgmgr.run_command import run_command
from pkgmgr.save_user_config import save_user_config
from pkgmgr.show_config import show_config
from pkgmgr.status_repos import status_repos
from pkgmgr.update_repos import update_repos
# Commands proxied by package-manager
PROXY_COMMANDS = {
"git":[
"pull",
"push",
"diff",
"add",
"show",
"checkout",
"clone",
"reset",
"revert",
"rebase",
"commit"
],
"docker":[
"start",
"stop",
"build"
],
"docker compose":[
"up",
"down",
"exec",
"ps",
"restart",
]
}
class SortedSubParsersAction(argparse._SubParsersAction):
def add_parser(self, name, **kwargs):
parser = super().add_parser(name, **kwargs)
# Sort the list of subparsers each time one is added
self._choices_actions.sort(key=lambda a: a.dest)
return parser
# Main program.
def main() -> None:
CONFIG_MERGED = load_config(USER_CONFIG_PATH)
REPOSITORIES_BASE_DIR = os.path.expanduser(CONFIG_MERGED["directories"]["repositories"])
ALL_REPOSITORIES = CONFIG_MERGED["repositories"]
BINARIES_DIRECTORY = os.path.expanduser(CONFIG_MERGED["directories"]["binaries"])
description_text = """\
\033[1;32mPackage Manager 🤖📦\033[0m
\033[3mKevin's Package Manager is a multi-repository, multi-package, and multi-format
development tool crafted by and designed for:\033[0m
\033[1;34mKevin Veen-Birkenbach\033[0m
\033[4mhttps://www.veen.world/\033[0m
\033[1mOverview:\033[0m
A powerful toolchain that unifies and automates workflows across heterogeneous
project ecosystems. pkgmgr is not only a package manager — it is a full
developer-oriented orchestration tool.
It automatically detects, merges, and processes metadata from multiple
dependency formats, including:
\033[1;33mPython:\033[0m pyproject.toml, requirements.txt
\033[1;33mNix:\033[0m flake.nix
\033[1;33mArch Linux:\033[0m PKGBUILD
\033[1;33mAnsible:\033[0m requirements.yml
\033[1;33mpkgmgr-native:\033[0m pkgmgr.yml
This allows pkgmgr to perform installation, updates, verification, dependency
resolution, and synchronization across complex multi-repo environments — with a
single unified command-line interface.
\033[1mDeveloper Tools:\033[0m
pkgmgr includes an integrated toolbox to enhance daily development workflows:
\033[1;33mVS Code integration:\033[0m Auto-generate and open multi-repo workspaces
\033[1;33mTerminal integration:\033[0m Open repositories in new GNOME Terminal tabs
\033[1;33mExplorer integration:\033[0m Open repositories in your file manager
\033[1;33mRelease automation:\033[0m Version bumping, changelog updates, and tagging
\033[1;33mBatch operations:\033[0m Execute shell commands across multiple repositories
\033[1;33mGit/Docker/Make wrappers:\033[0m Unified command proxying for many tools
\033[1mCapabilities:\033[0m
• Clone, pull, verify, update, and manage many repositories at once
• Resolve dependencies across languages and ecosystems
• Standardize install/update workflows
• Create symbolic executable wrappers for any project
• Merge configuration from default + user config layers
Use pkgmgr as both a robust package management framework and a versatile
development orchestration tool.
For detailed help on each command, use:
\033[1mpkgmgr <command> --help\033[0m
"""
parser = argparse.ArgumentParser(description=description_text,formatter_class=argparse.RawTextHelpFormatter)
subparsers = parser.add_subparsers(dest="command", help="Subcommands", action=SortedSubParsersAction)
def add_identifier_arguments(subparser):
subparser.add_argument(
"identifiers",
nargs="*",
help="Identifier(s) for repositories. Default: Repository of current folder.",
)
subparser.add_argument(
"--all",
action="store_true",
default=False,
help="Apply the subcommand to all repositories in the config. Some subcommands ask for confirmation. If you want to give this confirmation for all repositories, pipe 'yes'. E.g: yes | pkgmgr {subcommand} --all"
)
subparser.add_argument("--preview", action="store_true", help="Preview changes without executing commands")
subparser.add_argument("--list", action="store_true", help="List affected repositories (with preview or status)")
subparser.add_argument("-a", "--args", nargs=argparse.REMAINDER, dest="extra_args", help="Additional parameters to be attached.",default=[])
def add_install_update_arguments(subparser):
add_identifier_arguments(subparser)
subparser.add_argument(
"-q",
"--quiet",
action="store_true",
help="Suppress warnings and info messages",
)
subparser.add_argument(
"--no-verification",
action="store_true",
default=False,
help="Disable verification via commit/gpg",
)
subparser.add_argument(
"--dependencies",
action="store_true",
help="Also pull and update dependencies",
)
subparser.add_argument(
"--clone-mode",
choices=["ssh", "https", "shallow"],
default="ssh",
help="Specify the clone mode: ssh, https, or shallow (HTTPS shallow clone; default: ssh)",
)
install_parser = subparsers.add_parser("install", help="Setup repository/repositories alias links to executables")
add_install_update_arguments(install_parser)
update_parser = subparsers.add_parser("update", help="Update (pull + install) repository/repositories")
add_install_update_arguments(update_parser)
update_parser.add_argument("--system", action="store_true", help="Include system update commands")
deinstall_parser = subparsers.add_parser("deinstall", help="Remove alias links to repository/repositories")
add_identifier_arguments(deinstall_parser)
delete_parser = subparsers.add_parser("delete", help="Delete repository/repositories alias links to executables")
add_identifier_arguments(delete_parser)
# Add the 'create' subcommand (with existing identifier arguments)
create_parser = subparsers.add_parser(
"create",
help="Create new repository entries: add them to the config if not already present, initialize the local repository, and push remotely if --remote is set."
)
# Reuse the common identifier arguments
add_identifier_arguments(create_parser)
create_parser.add_argument(
"--remote",
action="store_true",
help="If set, add the remote and push the initial commit."
)
status_parser = subparsers.add_parser("status", help="Show status for repository/repositories or system")
add_identifier_arguments(status_parser)
status_parser.add_argument("--system", action="store_true", help="Show system status")
config_parser = subparsers.add_parser("config", help="Manage configuration")
config_subparsers = config_parser.add_subparsers(dest="subcommand", help="Config subcommands", required=True)
config_show = config_subparsers.add_parser("show", help="Show configuration")
add_identifier_arguments(config_show)
config_add = config_subparsers.add_parser("add", help="Interactively add a new repository entry")
config_edit = config_subparsers.add_parser("edit", help="Edit configuration file with nano")
config_init_parser = config_subparsers.add_parser("init", help="Initialize user configuration by scanning the base directory")
config_delete = config_subparsers.add_parser("delete", help="Delete repository entry from user config")
add_identifier_arguments(config_delete)
config_ignore = config_subparsers.add_parser("ignore", help="Set ignore flag for repository entries in user config")
add_identifier_arguments(config_ignore)
config_ignore.add_argument("--set", choices=["true", "false"], required=True, help="Set ignore to true or false")
path_parser = subparsers.add_parser("path", help="Print the path(s) of repository/repositories")
add_identifier_arguments(path_parser)
explore_parser = subparsers.add_parser("explore", help="Open repository in Nautilus file manager")
add_identifier_arguments(explore_parser)
terminal_parser = subparsers.add_parser("terminal", help="Open repository in a new GNOME Terminal tab")
add_identifier_arguments(terminal_parser)
release_parser = subparsers.add_parser(
"release",
help="Create a release for repository/ies by incrementing version and updating the changelog."
)
release_parser.add_argument(
"release_type",
choices=["major", "minor", "patch"],
help="Type of version increment for the release (major, minor, patch)."
)
release_parser.add_argument(
"-m", "--message",
default="",
help="Optional release message to add to the changelog and tag."
)
add_identifier_arguments(release_parser)
code_parser = subparsers.add_parser("code", help="Open repository workspace with VS Code")
add_identifier_arguments(code_parser)
list_parser = subparsers.add_parser("list", help="List all repositories with details and status")
list_parser.add_argument("--search", default="", help="Filter repositories that contain the given string")
list_parser.add_argument("--status", type=str, default="", help="Filter repositories by status (case insensitive)")
# Add the subcommand parser for "shell"
shell_parser = subparsers.add_parser("shell", help="Execute a shell command in each repository")
add_identifier_arguments(shell_parser)
shell_parser.add_argument("-c", "--command", nargs=argparse.REMAINDER, dest="shell_command", help="The shell command (and its arguments) to execute in each repository",default=[])
make_parser = subparsers.add_parser("make", help="Executes make commands")
add_identifier_arguments(make_parser)
make_subparsers = make_parser.add_subparsers(dest="subcommand", help="Make subcommands", required=True)
make_install = make_subparsers.add_parser("install", help="Executes the make install command")
add_identifier_arguments(make_install)
make_deinstall = make_subparsers.add_parser("deinstall", help="Executes the make deinstall command")
proxy_command_parsers = {}
for command, subcommands in PROXY_COMMANDS.items():
for subcommand in subcommands:
proxy_command_parsers[f"{command}_{subcommand}"] = subparsers.add_parser(
subcommand,
help=f"Proxies '{command} {subcommand}' to repository/ies",
description=f"Executes '{command} {subcommand}' for the identified repos.\nTo recieve more help execute '{command} {subcommand} --help'",
formatter_class=argparse.RawTextHelpFormatter
)
if subcommand in ["pull", "clone"]:
proxy_command_parsers[f"{command}_{subcommand}"].add_argument(
"--no-verification",
action="store_true",
default=False,
help="Disable verification via commit/gpg",
)
if subcommand == "clone":
proxy_command_parsers[f"{command}_{subcommand}"].add_argument(
"--clone-mode",
choices=["ssh", "https", "shallow"],
default="ssh",
help="Specify the clone mode: ssh, https, or shallow (HTTPS shallow clone; default: ssh)",
)
add_identifier_arguments(proxy_command_parsers[f"{command}_{subcommand}"])
args = parser.parse_args()
# All
if args.command and not args.command in ["config","list","create"]:
selected = get_selected_repos(args.all,ALL_REPOSITORIES,args.identifiers)
for command, subcommands in PROXY_COMMANDS.items():
for subcommand in subcommands:
if args.command == subcommand:
if args.command == "clone":
clone_repos(
selected,
REPOSITORIES_BASE_DIR,
ALL_REPOSITORIES,
args.preview,
args.no_verification,
args.clone_mode
)
elif args.command == "pull":
from pkgmgr.pull_with_verification import pull_with_verification
pull_with_verification(
selected,
REPOSITORIES_BASE_DIR,
ALL_REPOSITORIES,
args.extra_args,
args.no_verification,
args.preview
)
else:
exec_proxy_command(command,selected, REPOSITORIES_BASE_DIR, ALL_REPOSITORIES, args.command, args.extra_args, args.preview)
exit(0)
if args.command in ["make"]:
exec_proxy_command(args.command,selected, REPOSITORIES_BASE_DIR, ALL_REPOSITORIES, args.subcommand, args.extra_args, args.preview)
exit(0)
# Dispatch commands.
if args.command == "install":
install_repos(
selected,
REPOSITORIES_BASE_DIR,
BINARIES_DIRECTORY,
ALL_REPOSITORIES,
args.no_verification,
args.preview,
args.quiet,
args.clone_mode,
args.dependencies,
)
elif args.command == "create":
from pkgmgr.create_repo import create_repo
# If no identifiers are provided, you can decide to either use the repository of the current folder
# or prompt the user to supply at least one identifier.
if not args.identifiers:
print("No identifiers provided. Please specify at least one identifier in the format provider/account/repository.")
sys.exit(1)
else:
selected = get_selected_repos(True,ALL_REPOSITORIES,None)
for identifier in args.identifiers:
create_repo(identifier, CONFIG_MERGED, USER_CONFIG_PATH, BINARIES_DIRECTORY, remote=args.remote, preview=args.preview)
elif args.command == "list":
list_repositories(ALL_REPOSITORIES, REPOSITORIES_BASE_DIR, BINARIES_DIRECTORY, search_filter=args.search, status_filter=args.status)
elif args.command == "deinstall":
deinstall_repos(selected,REPOSITORIES_BASE_DIR, BINARIES_DIRECTORY, ALL_REPOSITORIES, preview=args.preview)
elif args.command == "delete":
delete_repos(selected,REPOSITORIES_BASE_DIR, ALL_REPOSITORIES, preview=args.preview)
elif args.command == "update":
update_repos(
selected,
REPOSITORIES_BASE_DIR,
BINARIES_DIRECTORY,
ALL_REPOSITORIES,
args.no_verification,
args.system,
args.preview,
args.quiet,
args.dependencies,
args.clone_mode
)
elif args.command == "release":
if not selected:
print("No repositories selected for release.")
exit(1)
# Import the release function from pkgmgr/release.py
from pkgmgr import release as rel
# Save the original working directory.
original_dir = os.getcwd()
for repo in selected:
# Determine the repository directory
repo_dir = repo.get("directory")
if not repo_dir:
from pkgmgr.get_repo_dir import get_repo_dir
repo_dir = get_repo_dir(REPOSITORIES_BASE_DIR, repo)
# Dynamically determine the file paths for pyproject.toml and CHANGELOG.md.
pyproject_path = os.path.join(repo_dir, "pyproject.toml")
changelog_path = os.path.join(repo_dir, "CHANGELOG.md")
print(f"Releasing repository '{repo.get('repository')}' in '{repo_dir}'...")
# Change into the repository directory so Git commands run in the right context.
os.chdir(repo_dir)
# Call the release function with the proper parameters.
rel.release(
pyproject_path=pyproject_path,
changelog_path=changelog_path,
release_type=args.release_type,
message=args.message
)
# Change back to the original working directory.
os.chdir(original_dir)
elif args.command == "status":
status_repos(selected,REPOSITORIES_BASE_DIR, ALL_REPOSITORIES, args.extra_args, list_only=args.list, system_status=args.system, preview=args.preview)
elif args.command == "explore":
for repository in selected:
run_command(f"nautilus {repository['directory']} & disown")
elif args.command == "code":
if not selected:
print("No repositories selected.")
else:
identifiers = [get_repo_identifier(repo, ALL_REPOSITORIES) for repo in selected]
sorted_identifiers = sorted(identifiers)
workspace_name = "_".join(sorted_identifiers) + ".code-workspace"
workspaces_dir = os.path.expanduser(CONFIG_MERGED.get("directories").get("workspaces"))
os.makedirs(workspaces_dir, exist_ok=True)
workspace_file = os.path.join(workspaces_dir, workspace_name)
folders = []
for repository in selected:
folders.append({"path": repository["directory"]})
workspace_data = {
"folders": folders,
"settings": {}
}
if not os.path.exists(workspace_file):
with open(workspace_file, "w") as f:
json.dump(workspace_data, f, indent=4)
print(f"Created workspace file: {workspace_file}")
else:
print(f"Using existing workspace file: {workspace_file}")
run_command(f'code "{workspace_file}"')
elif args.command == "terminal":
for repository in selected:
run_command(f'gnome-terminal --tab --working-directory="{repository["directory"]}"')
elif args.command == "path":
for repository in selected:
print(repository["directory"])
elif args.command == "shell":
if not args.shell_command:
print("No shell command specified.")
exit(2)
# Join the provided shell command parts into one string.
command_to_run = " ".join(args.shell_command)
for repository in selected:
print(f"Executing in '{repository['directory']}': {command_to_run}")
run_command(command_to_run, cwd=repository["directory"], preview=args.preview)
elif args.command == "config":
if args.subcommand == "show":
if args.all or (not args.identifiers):
show_config([], USER_CONFIG_PATH, full_config=True)
else:
selected = resolve_repos(args.identifiers, ALL_REPOSITORIES)
if selected:
show_config(selected, USER_CONFIG_PATH, full_config=False)
elif args.subcommand == "add":
interactive_add(CONFIG_MERGED,USER_CONFIG_PATH)
elif args.subcommand == "edit":
"""Open the user configuration file in nano."""
run_command(f"nano {USER_CONFIG_PATH}")
elif args.subcommand == "init":
if os.path.exists(USER_CONFIG_PATH):
with open(USER_CONFIG_PATH, 'r') as f:
user_config = yaml.safe_load(f) or {}
else:
user_config = {"repositories": []}
config_init(user_config, CONFIG_MERGED, BINARIES_DIRECTORY, USER_CONFIG_PATH)
elif args.subcommand == "delete":
# Load user config from USER_CONFIG_PATH.
if os.path.exists(USER_CONFIG_PATH):
with open(USER_CONFIG_PATH, 'r') as f:
user_config = yaml.safe_load(f) or {"repositories": []}
else:
user_config = {"repositories": []}
if args.all or not args.identifiers:
print("You must specify identifiers to delete.")
else:
to_delete = resolve_repos(args.identifiers, user_config.get("repositories", []))
new_repos = [entry for entry in user_config.get("repositories", []) if entry not in to_delete]
user_config["repositories"] = new_repos
save_user_config(user_config,USER_CONFIG_PATH)
print(f"Deleted {len(to_delete)} entries from user config.")
elif args.subcommand == "ignore":
# Load user config from USER_CONFIG_PATH.
if os.path.exists(USER_CONFIG_PATH):
with open(USER_CONFIG_PATH, 'r') as f:
user_config = yaml.safe_load(f) or {"repositories": []}
else:
user_config = {"repositories": []}
if args.all or not args.identifiers:
print("You must specify identifiers to modify ignore flag.")
else:
to_modify = resolve_repos(args.identifiers, user_config.get("repositories", []))
for entry in user_config["repositories"]:
key = (entry.get("provider"), entry.get("account"), entry.get("repository"))
for mod in to_modify:
mod_key = (mod.get("provider"), mod.get("account"), mod.get("repository"))
if key == mod_key:
entry["ignore"] = (args.set == "true")
print(f"Set ignore for {key} to {entry['ignore']}")
save_user_config(user_config,USER_CONFIG_PATH)
else:
parser.print_help()

View File

@@ -1,8 +1,10 @@
import os
import subprocess
import sys
import yaml
import tempfile
import shutil
import yaml
from pkgmgr.get_repo_identifier import get_repo_identifier
from pkgmgr.get_repo_dir import get_repo_dir
from pkgmgr.create_ink import create_ink
@@ -10,34 +12,221 @@ from pkgmgr.run_command import run_command
from pkgmgr.verify import verify_repository
from pkgmgr.clone_repos import clone_repos
def install_repos(
selected_repos,
repositories_base_dir,
bin_dir,
all_repos,
no_verification,
preview,
quiet,
clone_mode: str,
update_dependencies: bool
):
def _extract_pkgbuild_array(repo_dir: str, var_name: str) -> list:
"""
Install repositories by creating symbolic links, running setup commands, and
installing additional packages if a requirements.yml or requirements.txt file is found.
Extract a Bash array (depends/makedepends) from PKGBUILD using bash itself.
Returns a list of package names or an empty list on error.
"""
pkgbuild_path = os.path.join(repo_dir, "PKGBUILD")
if not os.path.exists(pkgbuild_path):
return []
script = f'source PKGBUILD >/dev/null 2>&1; printf "%s\\n" "${{{var_name}[@]}}"'
try:
output = subprocess.check_output(
["bash", "-lc", script],
cwd=repo_dir,
text=True,
)
except Exception:
return []
return [line.strip() for line in output.splitlines() if line.strip()]
def _install_arch_dependencies_from_pkgbuild(repo_dir: str, preview: bool) -> None:
"""
If PKGBUILD exists and pacman is available, install depends + makedepends
via pacman.
"""
if shutil.which("pacman") is None:
return
pkgbuild_path = os.path.join(repo_dir, "PKGBUILD")
if not os.path.exists(pkgbuild_path):
return
depends = _extract_pkgbuild_array(repo_dir, "depends")
makedepends = _extract_pkgbuild_array(repo_dir, "makedepends")
all_pkgs = depends + makedepends
if not all_pkgs:
return
cmd = "sudo pacman -S --noconfirm " + " ".join(all_pkgs)
run_command(cmd, preview=preview)
def _install_nix_flake_profile(repo_dir: str, preview: bool) -> None:
"""
If flake.nix exists and 'nix' is available, try to install a profile
from the flake. Convention: try .#pkgmgr, then .#default.
"""
flake_path = os.path.join(repo_dir, "flake.nix")
if not os.path.exists(flake_path):
return
if shutil.which("nix") is None:
print("Warning: flake.nix found but 'nix' command not available. Skipping flake setup.")
return
print("Nix flake detected, attempting to install profile output...")
for output in ("pkgmgr", "default"):
cmd = f"nix profile install {repo_dir}#{output}"
try:
run_command(cmd, preview=preview)
print(f"Nix flake output '{output}' successfully installed.")
break
except SystemExit as e:
print(f"[Warning] Failed to install Nix flake output '{output}': {e}")
def _install_pkgmgr_dependencies_from_manifest(
repo_dir: str,
no_verification: bool,
update_dependencies: bool,
clone_mode: str,
preview: bool,
) -> None:
"""
Read pkgmgr.yml (if present) and install referenced pkgmgr repository
dependencies.
Expected format:
version: 1
author: "..."
url: "..."
description: "..."
dependencies:
- repository: github:user/repo
version: main
reason: "Optional description"
"""
manifest_path = os.path.join(repo_dir, "pkgmgr.yml")
if not os.path.exists(manifest_path):
return
try:
with open(manifest_path, "r", encoding="utf-8") as f:
manifest = yaml.safe_load(f) or {}
except Exception as e:
print(f"Error loading pkgmgr.yml in '{repo_dir}': {e}")
return
dependencies = manifest.get("dependencies", []) or []
if not isinstance(dependencies, list) or not dependencies:
return
# Optional: show basic metadata (author/url/description) if present
author = manifest.get("author")
url = manifest.get("url")
description = manifest.get("description")
if not preview:
print("pkgmgr manifest detected:")
if author:
print(f" author: {author}")
if url:
print(f" url: {url}")
if description:
print(f" description: {description}")
dep_repo_ids = []
for dep in dependencies:
if not isinstance(dep, dict):
continue
repo_id = dep.get("repository")
if repo_id:
dep_repo_ids.append(str(repo_id))
# Optionally: update (pull) dependencies before installing
if update_dependencies and dep_repo_ids:
cmd_pull = "pkgmgr pull " + " ".join(dep_repo_ids)
try:
run_command(cmd_pull, preview=preview)
except SystemExit as e:
print(f"Warning: 'pkgmgr pull' for dependencies failed (exit code {e}).")
# Install dependencies one by one
for dep in dependencies:
if not isinstance(dep, dict):
continue
repo_id = dep.get("repository")
if not repo_id:
continue
version = dep.get("version")
reason = dep.get("reason")
if reason and not preview:
print(f"Installing dependency {repo_id}: {reason}")
else:
print(f"Installing dependency {repo_id}...")
cmd = f"pkgmgr install {repo_id}"
if version:
cmd += f" --version {version}"
if no_verification:
cmd += " --no-verification"
if update_dependencies:
cmd += " --dependencies"
if clone_mode:
cmd += f" --clone-mode {clone_mode}"
try:
run_command(cmd, preview=preview)
except SystemExit as e:
print(f"[Warning] Failed to install dependency '{repo_id}': {e}")
def install_repos(
selected_repos,
repositories_base_dir,
bin_dir,
all_repos,
no_verification,
preview,
quiet,
clone_mode: str,
update_dependencies: bool,
):
"""
Install repositories by creating symbolic links and processing standard
manifest files (pkgmgr.yml, PKGBUILD, flake.nix, Ansible requirements,
Python manifests, Makefile).
"""
for repo in selected_repos:
repo_identifier = get_repo_identifier(repo, all_repos)
repo_dir = get_repo_dir(repositories_base_dir, repo)
if not os.path.exists(repo_dir):
print(f"Repository directory '{repo_dir}' does not exist. Cloning it now...")
# Pass the clone_mode parameter to clone_repos
clone_repos([repo], repositories_base_dir, all_repos, preview, no_verification, clone_mode)
clone_repos(
[repo],
repositories_base_dir,
all_repos,
preview,
no_verification,
clone_mode,
)
if not os.path.exists(repo_dir):
print(f"Cloning failed for repository {repo_identifier}. Skipping installation.")
continue
verified_info = repo.get("verified")
verified_ok, errors, commit_hash, signing_key = verify_repository(repo, repo_dir, mode="local", no_verification=no_verification)
verified_ok, errors, commit_hash, signing_key = verify_repository(
repo,
repo_dir,
mode="local",
no_verification=no_verification,
)
if not no_verification and verified_info and not verified_ok:
print(f"Warning: Verification failed for {repo_identifier}:")
@@ -49,98 +238,82 @@ def install_repos(
continue
# Create the symlink using create_ink.
create_ink(repo, repositories_base_dir, bin_dir, all_repos, quiet=quiet, preview=preview)
# Check if a requirements.yml file exists and install additional packages.
create_ink(
repo,
repositories_base_dir,
bin_dir,
all_repos,
quiet=quiet,
preview=preview,
)
# 1) pkgmgr.yml (pkgmgr-internal manifest for other repositories)
_install_pkgmgr_dependencies_from_manifest(
repo_dir=repo_dir,
no_verification=no_verification,
update_dependencies=update_dependencies,
clone_mode=clone_mode,
preview=preview,
)
# 2) Arch: PKGBUILD (depends/makedepends)
_install_arch_dependencies_from_pkgbuild(repo_dir, preview=preview)
# 3) Nix: flake.nix
_install_nix_flake_profile(repo_dir, preview=preview)
# 4) Ansible: requirements.yml (only collections/roles)
req_file = os.path.join(repo_dir, "requirements.yml")
if os.path.exists(req_file):
try:
with open(req_file, "r") as f:
requirements = yaml.safe_load(f)
with open(req_file, "r", encoding="utf-8") as f:
requirements = yaml.safe_load(f) or {}
except Exception as e:
print(f"Error loading requirements.yml in {repo_identifier}: {e}")
continue # Skip to next repository if error occurs
if requirements:
# Install pacman packages if defined.
if "pacman" in requirements:
pacman_packages = requirements["pacman"]
if pacman_packages:
cmd = "sudo pacman -S --noconfirm " + " ".join(pacman_packages)
run_command(cmd, preview=preview)
# Install yay packages if defined.
if "yay" in requirements:
yay_packages = requirements["yay"]
if yay_packages:
cmd = "sudo -u aur_builder yay -S --noconfirm " + " ".join(yay_packages)
run_command(cmd, preview=preview)
# Install pkgmgr packages if defined.
if "pkgmgr" in requirements:
pkgmgr_packages = requirements["pkgmgr"]
if pkgmgr_packages:
if update_dependencies:
cmd_pull = "pkgmgr pull " + " ".join(pkgmgr_packages)
try:
run_command(cmd_pull, preview=preview)
except SystemExit as e:
print(f"Warning: 'pkgmgr pull' command failed (exit code {e}). Trying fallback clone...")
cmd_clone = "pkgmgr clone " + " ".join(pkgmgr_packages)
run_command(cmd_clone, preview=preview)
cmd = "pkgmgr install " + " ".join(pkgmgr_packages)
requirements = None
if no_verification:
cmd += " --no-verification"
if update_dependencies:
cmd += " --dependencies"
if clone_mode:
cmd += f" --clone-mode {clone_mode}"
run_command(cmd, preview=preview)
# Install pip packages if defined.
if "pip" in requirements:
pip_packages = requirements["pip"]
if pip_packages:
cmd = "python3 -m pip install " + " ".join(pip_packages)
run_command(cmd, preview=preview)
# Check if the requirements contain either 'collections' or 'roles'
if requirements and isinstance(requirements, dict):
if "collections" in requirements or "roles" in requirements:
print(f"Ansible dependencies found in {repo_identifier}, installing...")
# Build a new dictionary that only contains the Ansible dependencies
ansible_requirements = {}
if "collections" in requirements:
ansible_requirements["collections"] = requirements["collections"]
if "roles" in requirements:
ansible_requirements["roles"] = requirements["roles"]
# Write the ansible requirements to a temporary file.
with tempfile.NamedTemporaryFile(mode='w', suffix='.yml', delete=False) as tmp:
with tempfile.NamedTemporaryFile(
mode="w",
suffix=".yml",
delete=False,
) as tmp:
yaml.dump(ansible_requirements, tmp, default_flow_style=False)
tmp_filename = tmp.name
# Install Ansible collections if defined.
if "collections" in ansible_requirements:
print(f"Ansible collections found in {repo_identifier}, installing...")
cmd = f"ansible-galaxy collection install -r {tmp_filename}"
run_command(cmd, cwd=repo_dir, preview=preview)
# Install Ansible roles if defined.
if "roles" in ansible_requirements:
print(f"Ansible roles found in {repo_identifier}, installing...")
cmd = f"ansible-galaxy role install -r {tmp_filename}"
run_command(cmd, cwd=repo_dir, preview=preview)
# Check if a requirements.txt file exists and install Python packages.
# 5) Python: pyproject.toml (modern) / requirements.txt (classic)
pyproject_path = os.path.join(repo_dir, "pyproject.toml")
if os.path.exists(pyproject_path):
print(f"pyproject.toml found in {repo_identifier}, installing Python project...")
cmd = "~/.venvs/pkgmgr/bin/pip install ."
run_command(cmd, cwd=repo_dir, preview=preview)
req_txt_file = os.path.join(repo_dir, "requirements.txt")
if os.path.exists(req_txt_file):
print(f"requirements.txt found in {repo_identifier}, installing Python dependencies...")
cmd = "~/.venvs/pkgmgr/bin/pip install -r requirements.txt"
run_command(cmd, cwd=repo_dir, preview=preview)
# Check if a Makefile exists and run make.
# 6) Makefile: make install (if present)
makefile_path = os.path.join(repo_dir, "Makefile")
if os.path.exists(makefile_path):
cmd = "make install"

View File

@@ -1,13 +1,44 @@
import sys
import shutil
from .exec_proxy_command import exec_proxy_command
from .run_command import run_command
from .get_repo_identifier import get_repo_identifier
def status_repos(selected_repos, repositories_base_dir, all_repos, extra_args, list_only=False, system_status=False, preview=False):
def status_repos(
selected_repos,
repositories_base_dir,
all_repos,
extra_args,
list_only: bool = False,
system_status: bool = False,
preview: bool = False,
):
if system_status:
print("System status:")
# Arch / AUR updates (if yay / aur_builder is configured)
run_command("sudo -u aur_builder yay -Qu --noconfirm", preview=preview)
# Nix profile status (if Nix is available)
if shutil.which("nix") is not None:
print("\nNix profile status:")
try:
run_command("nix profile list", preview=preview)
except SystemExit as e:
print(f"[Warning] Failed to query Nix profiles: {e}")
if list_only:
for repo in selected_repos:
print(get_repo_identifier(repo, all_repos))
else:
exec_proxy_command('git',selected_repos, repositories_base_dir, all_repos, "status", extra_args, preview)
exec_proxy_command(
"git",
selected_repos,
repositories_base_dir,
all_repos,
"status",
extra_args,
preview,
)

View File

@@ -1,7 +1,10 @@
import sys
import shutil
from pkgmgr.pull_with_verification import pull_with_verification
from pkgmgr.install_repos import install_repos
def update_repos(
selected_repos,
repositories_base_dir,
@@ -10,9 +13,10 @@ def update_repos(
no_verification,
system_update,
preview: bool,
quiet: bool,
quiet: bool,
update_dependencies: bool,
clone_mode: str):
clone_mode: str,
):
"""
Update repositories by pulling latest changes and installing them.
@@ -29,12 +33,12 @@ def update_repos(
- clone_mode: Method to clone repositories (ssh or https).
"""
pull_with_verification(
selected_repos,
repositories_base_dir,
all_repos,
[],
no_verification,
preview
selected_repos,
repositories_base_dir,
all_repos,
[],
no_verification,
preview,
)
install_repos(
@@ -46,10 +50,19 @@ def update_repos(
preview,
quiet,
clone_mode,
update_dependencies
update_dependencies,
)
if system_update:
from pkgmgr.run_command import run_command
# Nix: upgrade all profile entries (if Nix is available)
if shutil.which("nix") is not None:
try:
run_command("nix profile upgrade '.*'", preview=preview)
except SystemExit as e:
print(f"[Warning] 'nix profile upgrade' failed: {e}")
# Arch / AUR system update
run_command("sudo -u aur_builder yay -Syu --noconfirm", preview=preview)
run_command("sudo pacman -Syyu --noconfirm", preview=preview)
run_command("sudo pacman -Syyu --noconfirm", preview=preview)

35
pyproject.toml Normal file
View File

@@ -0,0 +1,35 @@
[build-system]
# This tells pip/build how to build your project (PEP 517).
requires = ["setuptools>=77", "wheel"]
build-backend = "setuptools.build_meta"
[project]
name = "package-manager"
version = "0.1.0"
description = "Kevin's Package Manager is a configurable Python tool to manage multiple repositories via Bash."
readme = "README.md"
requires-python = ">=3.10"
# Use a simple SPDX license string to avoid the deprecation warning.
license = "MIT"
authors = [
{ name = "Kevin Veen-Birkenbach", email = "info@veen.world" }
]
# Python runtime dependencies.
dependencies = [
"PyYAML"
]
[project.urls]
Homepage = "https://www.veen.world"
Repository = "https://github.com/kevinveenbirkenbach/package-manager"
Funding-GitHub-Sponsors = "https://github.com/sponsors/kevinveenbirkenbach"
Funding-Patreon = "https://www.patreon.com/c/kevinveenbirkenbach"
Funding-BuyMeACoffee = "https://buymeacoffee.com/kevinveenbirkenbach"
Funding-PayPal = "https://s.veen.world/paypaldonate"
[tool.setuptools]
packages = ["pkgmgr"]

View File

@@ -1 +1,4 @@
PyYAML
# Legacy file used only if pip still installs from requirements.txt.
# You may delete this file once you switch entirely to pyproject.toml.
PyYAML

9
requirements.yml Normal file
View File

@@ -0,0 +1,9 @@
# This file defines Ansible Galaxy dependencies for this repository.
# It is used by `ansible-galaxy install -r requirements.yml`
# to download required collections and roles.
#
# Since the package-manager project does not rely on Ansible,
# both sections are intentionally left empty.
collections: []
roles: []