505 lines
15 KiB
Python
505 lines
15 KiB
Python
import sys
|
|
import os
|
|
import json
|
|
import csv
|
|
import asyncio
|
|
import platform
|
|
import subprocess
|
|
from functools import lru_cache
|
|
from pathlib import Path
|
|
from typing import List, Dict, Coroutine, Union, Tuple
|
|
from collections import Counter, defaultdict
|
|
from concurrent.futures import ThreadPoolExecutor
|
|
import multiprocessing
|
|
|
|
from . import info
|
|
from . import common
|
|
|
|
|
|
MAX_INT = sys.maxsize
|
|
|
|
|
|
def get_relative_path(kid: os.PathLike, parent: str) -> Union[List[str], None]:
|
|
"""
|
|
Return the relative path depth if relative, otherwise None.
|
|
|
|
Both the `kid` and `parent` should be absolute paths
|
|
"""
|
|
if parent == "":
|
|
return None
|
|
|
|
p_kid = Path(kid)
|
|
# p_kid = Path(kid).resolve()
|
|
try:
|
|
p_rel = p_kid.relative_to(parent)
|
|
except ValueError:
|
|
return None
|
|
rel = str(p_rel).split(os.sep)
|
|
if rel == ["."]:
|
|
rel = []
|
|
return rel
|
|
|
|
|
|
@lru_cache()
|
|
def get_repos() -> Dict[str, Dict[str, str]]:
|
|
"""
|
|
Return a `dict` of repo name to repo absolute path and repo type
|
|
|
|
"""
|
|
path_file = common.get_config_fname("repos.csv")
|
|
repos = {}
|
|
if os.path.isfile(path_file) and os.stat(path_file).st_size > 0:
|
|
with open(path_file) as f:
|
|
rows = csv.DictReader(
|
|
f, ["path", "name", "type", "flags"], restval=""
|
|
) # it's actually a reader
|
|
repos = {
|
|
r["name"]: {
|
|
"path": r["path"],
|
|
"type": r["type"],
|
|
"flags": r["flags"].split(),
|
|
}
|
|
for r in rows
|
|
if is_git(r["path"], include_bare=True)
|
|
}
|
|
return repos
|
|
|
|
|
|
@lru_cache()
|
|
def get_context() -> Union[Path, None]:
|
|
"""
|
|
Return context file path, or None if not set. Note that if in auto context
|
|
mode, the return value is not auto.context but the resolved context,
|
|
which could be None.
|
|
|
|
"""
|
|
config_dir = Path(common.get_config_dir())
|
|
matches = list(config_dir.glob("*.context"))
|
|
if len(matches) > 1:
|
|
print("Cannot have multiple .context file")
|
|
sys.exit(1)
|
|
if not matches:
|
|
return None
|
|
ctx = matches[0]
|
|
if ctx.stem == "auto":
|
|
# The context is set to be the group with minimal distance to cwd
|
|
candidate = None
|
|
min_dist = MAX_INT
|
|
for gname, prop in get_groups().items():
|
|
rel = get_relative_path(Path.cwd(), prop["path"])
|
|
if rel is None:
|
|
continue
|
|
d = len(rel)
|
|
if d < min_dist:
|
|
candidate = gname
|
|
min_dist = d
|
|
if not candidate:
|
|
ctx = None
|
|
else:
|
|
ctx = ctx.with_name(f"{candidate}.context")
|
|
return ctx
|
|
|
|
|
|
@lru_cache()
|
|
def get_groups() -> Dict[str, Dict[str, Union[str, List]]]:
|
|
"""
|
|
Return a `dict` of group name to group properties such as repo names and
|
|
group path.
|
|
"""
|
|
fname = common.get_config_fname("groups.csv")
|
|
groups = {}
|
|
repos = get_repos()
|
|
# Each line is: group-name:repo1 repo2 repo3:group-path
|
|
if os.path.isfile(fname) and os.stat(fname).st_size > 0:
|
|
with open(fname, "r") as f:
|
|
rows = csv.DictReader(
|
|
f, ["name", "repos", "path"], restval="", delimiter=":"
|
|
)
|
|
# filter out invalid repos
|
|
groups = {
|
|
r["name"]: {
|
|
"repos": [repo for repo in r["repos"].split() if repo in repos],
|
|
"path": r["path"],
|
|
}
|
|
for r in rows
|
|
}
|
|
return groups
|
|
|
|
|
|
def delete_repo_from_groups(repo: str, groups: Dict[str, Dict]) -> bool:
|
|
"""
|
|
Delete repo from groups
|
|
"""
|
|
deleted = False
|
|
for name in groups:
|
|
try:
|
|
groups[name]["repos"].remove(repo)
|
|
except ValueError as e:
|
|
pass
|
|
else:
|
|
deleted = True
|
|
return deleted
|
|
|
|
|
|
def replace_context(old: Union[Path, None], new: str):
|
|
""" """
|
|
auto = Path(common.get_config_dir()) / "auto.context"
|
|
if auto.exists():
|
|
old = auto
|
|
|
|
if new == "none": # delete
|
|
old and old.unlink()
|
|
elif old:
|
|
# ctx.rename(ctx.with_stem(new_name)) # only works in py3.9
|
|
old.rename(old.with_name(f"{new}.context"))
|
|
else:
|
|
Path(auto.with_name(f"{new}.context")).write_text("")
|
|
|
|
|
|
def get_choices() -> List[Union[str, None]]:
|
|
"""
|
|
Return all repo names, group names, and an additional empty list. The empty
|
|
list is added as a workaround of
|
|
argparse's problem with coexisting nargs='*' and choices.
|
|
See https://utcc.utoronto.ca/~cks/space/blog/python/ArgparseNargsChoicesLimitation
|
|
and
|
|
https://bugs.python.org/issue27227
|
|
"""
|
|
choices = list(get_repos())
|
|
choices.extend(get_groups())
|
|
choices.append([])
|
|
return choices
|
|
|
|
|
|
def is_submodule_repo(p: Path) -> bool:
|
|
""" """
|
|
if p.is_file() and ".git/modules" in p.read_text():
|
|
return True
|
|
return False
|
|
|
|
|
|
def is_git(path: str, include_bare=False, exclude_submodule=False) -> bool:
|
|
"""
|
|
Return True if the path is a git repo.
|
|
"""
|
|
if not os.path.exists(path):
|
|
return False
|
|
# An alternative is to call `git rev-parse --is-inside-work-tree`
|
|
# I don't see why that one is better yet.
|
|
# For a regular git repo, .git is a folder. For a worktree repo and
|
|
# submodule repo, .git is a file.
|
|
# A more reliable way to differentiable regular and worktree repos is to
|
|
# compare the result of `git rev-parse --git-dir` and
|
|
# `git rev-parse --git-common-dir`
|
|
loc = os.path.join(path, ".git")
|
|
# TODO: we can display the worktree repos in a different font.
|
|
if os.path.exists(loc):
|
|
if exclude_submodule and is_submodule_repo(Path(loc)):
|
|
return False
|
|
return True
|
|
if not include_bare:
|
|
return False
|
|
# detect bare repo
|
|
got = subprocess.run(
|
|
"git rev-parse --is-bare-repository".split(),
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.DEVNULL,
|
|
cwd=path,
|
|
)
|
|
if got.returncode == 0 and got.stdout == b"true\n":
|
|
return True
|
|
return False
|
|
|
|
|
|
def rename_repo(repos: Dict[str, Dict[str, str]], repo: str, new_name: str):
|
|
"""
|
|
Write new repo name to file
|
|
"""
|
|
if new_name in repos:
|
|
print(f"{new_name} is already in use!")
|
|
return
|
|
prop = repos[repo]
|
|
del repos[repo]
|
|
repos[new_name] = prop
|
|
write_to_repo_file(repos, "w")
|
|
|
|
groups = get_groups()
|
|
for g, values in groups.items():
|
|
members = values["repos"]
|
|
if repo in members:
|
|
members.remove(repo)
|
|
members.append(new_name)
|
|
groups[g]["repos"] = sorted(members)
|
|
write_to_groups_file(groups, "w")
|
|
|
|
|
|
def write_to_repo_file(repos: Dict[str, Dict[str, str]], mode: str):
|
|
"""
|
|
@param repos: each repo is {name: {properties}}
|
|
"""
|
|
# The 3rd column is repo type; unused field
|
|
data = [
|
|
(prop["path"], name, "", " ".join(prop["flags"]))
|
|
for name, prop in repos.items()
|
|
]
|
|
fname = common.get_config_fname("repos.csv")
|
|
os.makedirs(os.path.dirname(fname), exist_ok=True)
|
|
with open(fname, mode, newline="") as f:
|
|
writer = csv.writer(f, delimiter=",", quotechar='"', quoting=csv.QUOTE_MINIMAL)
|
|
writer.writerows(data)
|
|
|
|
|
|
# TODO: combine with the repo writer
|
|
def write_to_groups_file(groups: Dict[str, Dict], mode: str):
|
|
""" """
|
|
fname = common.get_config_fname("groups.csv")
|
|
os.makedirs(os.path.dirname(fname), exist_ok=True)
|
|
if not groups: # all groups are deleted
|
|
Path(fname).write_text("")
|
|
else:
|
|
# delete the group if there are no repos
|
|
for name in list(groups):
|
|
if not groups[name]["repos"]:
|
|
del groups[name]
|
|
with open(fname, mode, newline="") as f:
|
|
data = [
|
|
(group, " ".join(prop["repos"]), prop["path"])
|
|
for group, prop in groups.items()
|
|
]
|
|
writer = csv.writer(
|
|
f, delimiter=":", quotechar='"', quoting=csv.QUOTE_MINIMAL
|
|
)
|
|
writer.writerows(data)
|
|
|
|
|
|
def _make_name(
|
|
path: str, repos: Dict[str, Dict[str, str]], name_counts: Counter
|
|
) -> str:
|
|
"""
|
|
Given a new repo `path`, create a repo name. By default, basename is used.
|
|
If name collision exists, further include parent path name.
|
|
|
|
@param path: It should not be in `repos` and is absolute
|
|
"""
|
|
name = os.path.basename(os.path.normpath(path))
|
|
if name in repos or name_counts[name] > 1:
|
|
# path has no trailing /
|
|
par_name = os.path.basename(os.path.dirname(path))
|
|
return os.path.join(par_name, name)
|
|
return name
|
|
|
|
|
|
def add_repos(
|
|
repos: Dict[str, Dict[str, str]],
|
|
new_paths: List[str],
|
|
include_bare=False,
|
|
exclude_submodule=False,
|
|
dry_run=False,
|
|
) -> Dict[str, Dict[str, str]]:
|
|
"""
|
|
Write new repo paths to file; return the added repos.
|
|
|
|
@param repos: name -> path
|
|
"""
|
|
existing_paths = {prop["path"] for prop in repos.values()}
|
|
new_paths = {p for p in new_paths if is_git(p, include_bare, exclude_submodule)}
|
|
new_paths = new_paths - existing_paths
|
|
new_repos = {}
|
|
if new_paths:
|
|
print(f"Found {len(new_paths)} new repo(s).")
|
|
if dry_run:
|
|
for p in new_paths:
|
|
print(p)
|
|
return {}
|
|
name_counts = Counter(os.path.basename(os.path.normpath(p)) for p in new_paths)
|
|
new_repos = {
|
|
_make_name(path, repos, name_counts): {
|
|
"path": path,
|
|
"flags": "",
|
|
}
|
|
for path in new_paths
|
|
}
|
|
write_to_repo_file(new_repos, "a+")
|
|
else:
|
|
print("No new repos found!")
|
|
return new_repos
|
|
|
|
|
|
def _generate_dir_hash(repo_path: str, paths: List[str]) -> Tuple[Tuple[str, ...], str]:
|
|
"""
|
|
Return relative parent strings, and the parent head string
|
|
|
|
For example, if `repo_path` is /a/b/c/d/here, and one of `paths` is /a/b/
|
|
then return (b, c, d)
|
|
"""
|
|
for p in paths:
|
|
rel = get_relative_path(repo_path, p)[:-1]
|
|
if rel is not None:
|
|
break
|
|
else:
|
|
return (), ""
|
|
head, tail = os.path.split(p)
|
|
return (tail, *rel), head
|
|
|
|
|
|
def auto_group(repos: Dict[str, Dict[str, str]], paths: List[str]) -> Dict[str, Dict]:
|
|
"""
|
|
|
|
@params repos: repos to be grouped
|
|
"""
|
|
# FIXME: the upstream code should make sure that paths are all independent
|
|
# i.e., each repo should be contained in one and only one path
|
|
new_groups = defaultdict(dict)
|
|
for repo_name, prop in repos.items():
|
|
hash, head = _generate_dir_hash(prop["path"], paths)
|
|
if not hash:
|
|
continue
|
|
for i in range(1, len(hash) + 1):
|
|
group_name = "-".join(hash[:i])
|
|
prop = new_groups[group_name]
|
|
prop["path"] = os.path.join(head, *hash[:i])
|
|
if "repos" not in prop:
|
|
prop["repos"] = [repo_name]
|
|
else:
|
|
prop["repos"].append(repo_name)
|
|
# FIXME: need to make sure the new group names don't clash with old ones
|
|
# or repo names
|
|
return new_groups
|
|
|
|
|
|
async def run_async(repo_name: str, path: str, cmds: List[str]) -> Union[None, str]:
|
|
"""
|
|
Run `cmds` asynchronously in `path` directory. Return the `path` if
|
|
execution fails.
|
|
"""
|
|
# TODO: deprecated since 3.8, will be removed in 3.10
|
|
process = await asyncio.create_subprocess_exec(
|
|
*cmds,
|
|
stdin=asyncio.subprocess.DEVNULL,
|
|
stdout=asyncio.subprocess.PIPE,
|
|
stderr=asyncio.subprocess.PIPE,
|
|
start_new_session=True,
|
|
cwd=path,
|
|
)
|
|
stdout, stderr = await process.communicate()
|
|
for pipe in (stdout, stderr):
|
|
if pipe:
|
|
print(format_output(pipe.decode(), repo_name))
|
|
# The existence of stderr is not good indicator since git sometimes write
|
|
# to stderr even if the execution is successful, e.g. git fetch
|
|
if process.returncode != 0:
|
|
return path
|
|
|
|
|
|
def format_output(s: str, prefix: str):
|
|
"""
|
|
Prepends every line in given string with the given prefix.
|
|
"""
|
|
return "".join([f"{prefix}: {line}" for line in s.splitlines(keepends=True)])
|
|
|
|
|
|
def exec_async_tasks(tasks: List[Coroutine]) -> List[Union[None, str]]:
|
|
"""
|
|
Execute tasks asynchronously
|
|
"""
|
|
# TODO: asyncio API is nicer in python 3.7
|
|
if platform.system() == "Windows":
|
|
loop = asyncio.ProactorEventLoop()
|
|
asyncio.set_event_loop(loop)
|
|
else:
|
|
loop = asyncio.get_event_loop()
|
|
|
|
try:
|
|
errors = loop.run_until_complete(asyncio.gather(*tasks))
|
|
finally:
|
|
loop.close()
|
|
return errors
|
|
|
|
|
|
def describe(repos: Dict[str, Dict[str, str]], no_colors: bool = False) -> str:
|
|
"""
|
|
Return the status of all repos
|
|
"""
|
|
if repos:
|
|
truncator = info.Truncate()
|
|
name_width = len(max(repos, key=len)) + 1
|
|
funcs = info.get_info_funcs(no_colors=no_colors)
|
|
|
|
num_threads = min(multiprocessing.cpu_count(), len(repos))
|
|
with ThreadPoolExecutor(max_workers=num_threads) as executor:
|
|
for line in executor.map(
|
|
lambda name: f'{name:<{name_width}}{" ".join(f(repos[name], truncator) for f in funcs)}',
|
|
sorted(repos),
|
|
):
|
|
yield line
|
|
|
|
|
|
def get_cmds_from_files() -> Dict[str, Dict[str, str]]:
|
|
"""
|
|
Parse delegated git commands from default config file
|
|
and custom config file.
|
|
|
|
Example return
|
|
{
|
|
'branch': {'help': 'show local branches'},
|
|
'clean': {'cmd': 'clean -dfx',
|
|
'help': 'remove all untracked files/folders'},
|
|
}
|
|
"""
|
|
# default config file
|
|
fname = os.path.join(os.path.dirname(__file__), "cmds.json")
|
|
with open(fname, "r") as f:
|
|
cmds = json.load(f)
|
|
|
|
# custom config file
|
|
root = common.get_config_dir()
|
|
fname = os.path.join(root, "cmds.json")
|
|
custom_cmds = {}
|
|
if os.path.isfile(fname) and os.path.getsize(fname):
|
|
with open(fname, "r") as f:
|
|
custom_cmds = json.load(f)
|
|
|
|
# custom commands shadow default ones
|
|
cmds.update(custom_cmds)
|
|
return cmds
|
|
|
|
|
|
def parse_repos_and_rest(
|
|
input: List[str],
|
|
quote_mode=False,
|
|
) -> Tuple[Dict[str, Dict[str, str]], List[str]]:
|
|
"""
|
|
Parse gita input arguments
|
|
|
|
@return: repos and the rest (e.g., gita shell and super commands)
|
|
"""
|
|
i = 0
|
|
names = []
|
|
repos = get_repos()
|
|
groups = get_groups()
|
|
ctx = get_context()
|
|
for i, word in enumerate(input):
|
|
if word in repos or word in groups:
|
|
names.append(word)
|
|
else:
|
|
break
|
|
else: # all input is repos and groups, shift the index once more
|
|
if i is not None:
|
|
i += 1
|
|
if not names and ctx:
|
|
names = [ctx.stem]
|
|
if quote_mode and i + 1 != len(input):
|
|
print(input[i], "is not a repo or group")
|
|
sys.exit(2)
|
|
|
|
if names:
|
|
chosen = {}
|
|
for k in names:
|
|
if k in repos:
|
|
chosen[k] = repos[k]
|
|
if k in groups:
|
|
for r in groups[k]["repos"]:
|
|
chosen[r] = repos[r]
|
|
# if not set here, all repos are chosen
|
|
repos = chosen
|
|
return repos, input[i:]
|