# Copyright (c) 2003-2012 LOGILAB S.A. (Paris, FRANCE). # http://www.logilab.fr/ -- mailto:contact@logilab.fr # # This program is free software; you can redistribute it and/or modify it under # the terms of the GNU General Public License as published by the Free Software # Foundation; either version 2 of the License, or (at your option) any later # version. # # This program is distributed in the hope that it will be useful, but WITHOUT # ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS # FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program. If not, see . """helper functions and classes to ease hg revision graph building Based on graphlog's algorithm, with inspiration stolen to TortoiseHg revision grapher. """ import re try: from io import StringIO except ImportError: # Python 2 from cStringIO import StringIO import difflib from itertools import chain, count from time import strftime, localtime from functools import partial import collections from mercurial.node import nullrev from mercurial import patch, match, error import hgviewlib.hgpatches # force apply patches to mercurial from hgviewlib.hgpatches import mqsupport, phases, hiddenrevs from hgviewlib.util import (tounicode, tohg, isbfile, first_known_precursors,\ build_repo, allbranches, binary) from hgviewlib.config import HgConfig DATE_FMT = '%F %R' # match the end of the diff header, assuming that the following line # looks like "@@ -25,6 +23,5 @@" DIFFHEADERMATCHER = re.compile('^@@.+@@$', re.MULTILINE) def diff(repo, ctx1, ctx2=None, files=None): """ Compute the diff of ``files`` between the 2 contexts ``ctx1`` and ``ctx2``. :Note: context may be a changectx or a filectx. * If ``ctx2`` is None, the parent of ``ctx1`` is used. If ``ctx1`` is a file ctx, the parent is the first ancestor that contains modification on the given file * If ``files`` is None, return the diff for all files. """ if not getattr(ctx1, 'applied', True): #no diff vs ctx2 on unapplied patch return ''.join(chain(ctx1.filectx(tohg(fname)).data() for fname in files)) if ctx2 is None: ctx2 = ctx1.p1() if files is None: try: matchfn = match.always() except TypeError: # backward compat with mercurial < 5.0 matchfn = match.always(repo.root, repo.getcwd()) else: try: matchfn = match.exact([tohg(fname) for fname in files]) except TypeError: # backward compat with mercurial < 5.0: exact() takes at least 3 arguments (1 given) matchfn = match.exact(repo.root, repo.getcwd(), [tohg(fname) for fname in files]) # try/except for the sake of hg compatibility (API changes between # 1.0 and 1.1) diffopts = patch.diffopts(repo.ui) try: out = StringIO() patch.diff(repo, ctx2.node(), ctx1.node(), match=matchfn, fp=out, opts=diffopts) diffdata = out.getvalue() except: diffdata = b'\n'.join(patch.diff(repo, ctx2.node(), ctx1.node(), match=matchfn, opts=diffopts)) return tounicode(diffdata) def __get_parents(repo, rev, subset, cache): """ Return non-null parents of `rev`. Only revision in subset are taken in account. """ # we need to manage our own stack or we overflow python one. toproceed = collections.deque() if rev not in cache: toproceed.append(rev) else: final = cache[rev] while toproceed: current = toproceed.pop() if current is None: parents = [x.rev() for x in repo[None].parents() if x] else: parents = [x for x in repo.changelog.parentrevs(current) if x != nullrev] final = [] added = set() for p in parents: if p in subset: final.append(p) added.add(p) else: if p not in cache: # stack the parent and jump to next iteration toproceed.append(current) toproceed.append(p) break else: for p in cache[p]: if p not in added: final.append(p) added.add(p) else: # all parents in cache register the new one cache[current] = final return final def getlog(model, ctx, gnode): if ctx.rev() is not None: msg = tounicode(ctx.description()) if msg: msg = msg.splitlines()[0] else: msg = u"WORKING DIRECTORY (locally modified)" return msg def gettags(model, ctx, gnode=None): if ctx.rev() is None: return u"" mqtags = ['qbase', 'qtip', 'qparent'] tags = ctx.tags() if model.hide_mq_tags: tags = [t for t in tags if t not in mqtags] return u",".join(tounicode(t) for t in tags) def getdate(model, ctx, gnode): if not ctx.date(): return "" return strftime(DATE_FMT, localtime(int(ctx.date()[0]))) def ismerge(ctx): """ Return True if the changecontext ctx is a merge mode (should work with hg 1.0 and 1.2) """ if ctx: return len(ctx.parents()) == 2 and ctx.parents()[1] return False def _dirty_wc(repo): """return true if the working directory has changes""" return bool(sum(len(l) for l in repo.status())) # dirty working dir def _rev_order(phaserevs, rev): """return a sort key for changeset reordering (, rev) Freshness is: :0: public revision :1: mutable revision :42: working directory """ if rev is None: return (42, rev) return (phases.public != phaserevs[rev], rev) def _build_filter(start_rev, branch, follow, closed): """process filter rules and returns maching revisions as list If start_rev is None, started from working directory node If follow is True, only generated the subtree from the start_rev head. If branch is set, only generated the subtree for the given named branch. """ # selected range revset_args = [] if (start_rev is None): revset = 'all()' elif follow: revset = '::(%s)' % start_rev else: revset = ':(%s)' % start_rev # branch restriction if branch: # we add both changeset that belong to the branch and other merged. revset = '(%s and (branch(%%s) or parents(branch(%%s))))' % (revset) revset_args.extend([branch, branch]) # closed branch restriction if not closed: revset = '(%s and ::(head() - closed()))' % revset return revset, revset_args def _revset_revs(repo, revset, revset_args=(), view=b'visible'): """Use the fastest available method to get revs from a revset The view argument is used to control the available revision (see `repoview` in Mercurila core). We currently support "visible" and None as we have to implement backward compat manually. The revs are returned in ordered.""" revset = 'sort(%s)' % revset assert view in (b'visible', None), view if getattr(repo, 'filtered', None) is not None: # Mercurial 2.5 and above. repoview available, relying on it. if view is None: repo = repo.unfiltered() else: repo = repo.filtered(view) return repo.revs(revset, *revset_args) else: if getattr(repo, 'revs', None) is None: # pre 2.1 hg revs = [c.rev() for c in repo.set(revset, *revset_args)] else: revs = repo.revs(revset, *revset_args) excluded = () if view is None else hiddenrevs(repo) if excluded: revs = [r for r in revs if r not in excluded] return revs def revision_grapher(repo, revset_filter, start_rev=None, show_hidden=False, reorder=False, show_obsolete=False): """incremental revision grapher This generator function walks through the revision history from revision start_rev for each revision emits tuples with the following elements: - current revision - column of the current node in the set of ongoing edges - color of the node (?) - lines; a list of (col, next_col, color) indicating the edges between the current row and the next row - parent revisions of current revision """ # special handling of mq patches if show_hidden and start_rev == None and hasattr(repo, 'mq'): series = list(reversed(repo.mq.series)) for patchname in series: if not repo.mq.isapplied(patchname): yield (patchname, 0, 0, [(0, 0 ,0, False)], []) view = None if show_hidden else b'visible' included = _revset_revs(repo, *revset_filter, view=view) if getattr(included, 'append', None) is None: # _revset_revs returned a smartset (new in Mercurial 3.0) # translate it back to list for now included = list(included) # working directory if start_rev is None and _dirty_wc(repo): included.append(None) included.reverse() inc_set = set(included) # necessary for first_known_precursors # the second user of first_known_precursors can't really # reverse its logic. excluded = set(repo) - inc_set parentscache = {} # all known revs for this line. This is used to compute column index # it's combined with next_revs to compute how we must draw lines revs = [] levels = [] # a rev -> level mapping. # level are True for real relation (parent), # False for weak one (obsolete) rev_color = {} free_color = count(0) parent_func = partial(__get_parents, repo=repo, subset=inc_set, cache=parentscache) for curr_rev in included: # Compute revs and next_revs. if curr_rev not in revs: # rev not ancestor of already processed node # we add this new head to know revision revs.append(curr_rev) levels.append(True) rev_color[curr_rev] = curcolor = next(free_color) else: curcolor = rev_color[curr_rev] # copy known revisions for this line next_revs = revs[:] next_levels = levels[:] # Add parents to next_revs. parents = [(p, True) for p in parent_func(rev=curr_rev)] if show_obsolete: ctx = repo[curr_rev] for prec in first_known_precursors(ctx, excluded): parents.append((prec.rev(), False)) parents_to_add = [] max_levels = dict(zip(next_revs, next_levels)) for idx, (parent, level) in enumerate(parents): # could have been added by another children if parent not in next_revs: parents_to_add.append(parent) if idx == 0: # first parent inherit the color rev_color[parent] = curcolor else: # second don't rev_color[parent] = next(free_color) max_levels[parent] = level or max_levels.get(parent, False) # rev_index is also the column index rev_index = next_revs.index(curr_rev) # replace curr_rev by its parents. next_revs[rev_index:rev_index + 1] = parents_to_add next_levels = [max_levels[r] for r in next_revs] lines = [] for i, rev in enumerate(revs): if rev == curr_rev: # one or more line to parents targets = parents else: # single line to the same rev targets = [(rev, levels[i])] for trg, level in targets: color = rev_color[trg] lines.append((i, next_revs.index(trg), color, level)) yield (curr_rev, rev_index, curcolor, lines, parents) revs = next_revs levels = next_levels def filelog_grapher(repo, path): ''' Graph the ancestry of a single file (log). Deletions show up as breaks in the graph. ''' path = tohg(path) filerev = len(repo.file(path)) - 1 fctx = repo.filectx(path, fileid=filerev) rev = fctx.rev() flog = fctx.filelog() heads = [repo.filectx(path, fileid=flog.rev(x)).rev() for x in flog.heads()] assert rev in heads, (rev, heads) heads.remove(rev) revs = [] rev_color = {} nextcolor = 0 _paths = {} while rev >= 0: # Compute revs and next_revs if rev not in revs: revs.append(rev) rev_color[rev] = nextcolor ; nextcolor += 1 curcolor = rev_color[rev] index = revs.index(rev) next_revs = revs[:] # Add parents to next_revs fctx = repo.filectx(_paths.get(rev, path), changeid=rev) for pfctx in fctx.parents(): _paths[pfctx.rev()] = pfctx.path() parents = [pfctx.rev() for pfctx in fctx.parents()]# if f.path() == path] parents_to_add = [] for parent in parents: if parent not in next_revs: parents_to_add.append(parent) if len(parents) > 1: rev_color[parent] = nextcolor ; nextcolor += 1 else: rev_color[parent] = curcolor parents_to_add.sort() next_revs[index:index + 1] = parents_to_add lines = [] for i, nrev in enumerate(revs): if nrev in next_revs: color = rev_color[nrev] lines.append( (i, next_revs.index(nrev), color, True) ) elif nrev == rev: for parent in parents: color = rev_color[parent] lines.append( (i, next_revs.index(parent), color, True) ) pcrevs = [pfc.rev() for pfc in fctx.parents()] yield (fctx.rev(), index, curcolor, lines, pcrevs, tounicode(_paths.get(fctx.rev(), path))) revs = next_revs if revs: rev = max(revs) else: rev = -1 if heads and rev <= heads[-1]: rev = heads.pop() class GraphNode(object): """ Simple class to encapsulate e hg node in the revision graph. Does nothing but declaring attributes. """ def __init__(self, rev, xposition, color, lines, parents, ncols=None, extra=None): self.rev = rev self.x = xposition self.color = color if ncols is None: ncols = len(lines) self.cols = ncols if not parents: self.cols += 1 # root misses its own column self.parents = parents self.bottomlines = lines self.toplines = [] self.extra = extra class Graph(object): """ Graph object to ease hg repo navigation. The Graph object instantiate a `revision_grapher` generator, and provide a `fill` method to build the graph progressively. """ #@timeit def __init__(self, repo, grapher, maxfilesize=100000): self.maxfilesize = maxfilesize self.repo = repo self.maxlog = len(self.repo.changelog) self.grapher = grapher self.nodes = [] self.nodesdict = {} self.max_cols = 0 def build_nodes(self, nnodes=None, rev=None): """ Build up to `nnodes` more nodes in our graph, or build as many nodes required to reach `rev`. If both rev and nnodes are set, build as many nodes as required to reach rev plus nnodes more. """ if self.grapher is None: return False stopped = False mcol = [self.max_cols] for vnext in self.grapher: if vnext is None: continue nrev, xpos, color, lines, parents = vnext[:5] if isinstance(nrev, int) and nrev >= self.maxlog: continue gnode = GraphNode(nrev, xpos, color, lines, parents, extra=vnext[5:]) if self.nodes: gnode.toplines = self.nodes[-1].bottomlines self.nodes.append(gnode) self.nodesdict[nrev] = gnode mcol.append(gnode.cols) if rev is not None and nrev <= rev: rev = None # we reached rev, switching to nnodes counter if rev is None: if nnodes is not None: nnodes -= 1 if not nnodes: break else: break else: self.grapher = None stopped = True self.max_cols = max(mcol) return not stopped def isfilled(self): return self.grapher is None def fill(self, step=100): """ Return a generator that fills the graph by bursts of `step` more nodes at each iteration. """ while self.build_nodes(step): yield len(self) yield len(self) def __getitem__(self, idx): if isinstance(idx, slice): # XXX TODO: ensure nodes are built return self.nodes.__getitem__(idx) if idx >= len(self.nodes): # build as many graph nodes as required to answer the # requested idx self.build_nodes(idx) if idx > len(self): return self.nodes[-1] return self.nodes[idx] def __len__(self): # len(graph) is the number of actually built graph nodes return len(self.nodes) def index(self, rev): if len(self) == 0: # graph is empty, let's build some nodes self.build_nodes(10) if rev is not None and rev < self.nodes[-1].rev: self.build_nodes(self.nodes[-1].rev - rev) if rev in self.nodesdict: return self.nodes.index(self.nodesdict[rev]) return -1 def fileflags(self, filename, rev, _cache={}): """ Return a couple of flags ('=', '+', '-' or '?') depending on the nature of the diff for filename between rev and its parents. """ if rev not in _cache: ctx = self.repo[rev] _cache.clear() _cache[rev] = (ctx, [list(self.repo.status(p.node(), ctx.node()))[:5] for p in ctx.parents()]) ctx, allchanges = _cache[rev] flags = [] for changes in allchanges: # changes = modified, added, removed, deleted, unknown for flag, lst in zip(["=", "+", "-", "-", "?"], changes): if tohg(filename) in lst: if flag == "+": renamed = ctx.filectx(tohg(filename)).renamed() if renamed: flags.append(renamed) break flags.append(flag) break else: flags.append('') return flags def fileflag(self, filename, rev): """ Return a flag (see fileflags) between rev and its first parent (may be long) """ return self.fileflags(filename, rev)[0] def filename(self, rev): return self.nodesdict[rev].extra[0] def filedata(self, filename, rev, mode='diff', flag=None, maxfilesize=None): """XXX written under dubious encoding assumptions The modification flag is computed using *fileflag* if ``flag`` is None. """ # XXX This really begins to be a dirty mess... if maxfilesize is None: maxfilesize = self.maxfilesize data = "" if flag is None: flag = self.fileflag(filename, rev) ctx = self.repo[rev] filesize = 0 try: fctx = ctx.filectx(tohg(filename)) filesize = fctx.size() # compute size here to lookup data securely except (LookupError, OSError): fctx = None # may happen for renamed/removed files or mq patch ? if isbfile(filename): data = u"[bfile]\n" if fctx: data = fctx.data() data += u"footprint: %s\n" % tounicode(data) return "+", data if flag not in ('-', '?'): if fctx is None:# or fctx.node() is None: return '', None if maxfilesize >= 0 and filesize > maxfilesize: try: div = int(filesize).bit_length() // 10 sym = ('', 'K', 'M', 'G', 'T', 'E')[div] # more, really ??? val = int(filesize // (2 ** (div * 10))) except AttributeError: # py < 2.7 val = filesize sym = '' data = u"~%i%so" % (val, sym) return 'file too big', data if flag == "+" or mode == 'file': flag = '+' # return the whole file data = fctx.data() if binary(data): data = u"binary file" else: # tries to convert to unicode data = tounicode(data) elif flag == "=" or isinstance(mode, int): flag = "=" if isinstance(mode, int): parentctx = self.repo[mode] else: parentctx = self.repo[self._fileparent(fctx)] data = diff(self.repo, ctx, parentctx, files=[filename]) match = DIFFHEADERMATCHER.search(data) if match is not None: datastart = match.start() else: datastart = 0 data = data[datastart:] elif flag == '': data = u'' else: # file renamed oldname, node = flag newdata = tounicode(fctx.data()).splitlines() olddata = self.repo.filectx(oldname, fileid=node) olddata = tounicode(olddata.data()).splitlines() data = list(difflib.unified_diff(olddata, newdata, tounicode(oldname), filename))[2:] if data: flag = "=" else: data = newdata flag = "+" data = u'\n'.join(data) return flag, data def _fileparent(self, fctx): try: return fctx.p1().rev() except IndexError: # reach bottom return -1 def fileparent(self, filename, rev): return self._fileparent(self.repo[rev].filectx(tohg(filename))) class HgRepoListWalker(object): """ Graph object to ease hg repo revision tree drawing depending on user's configurations. """ _allcolumns = ('ID', 'Branch', 'Log', 'Author', 'Date', 'Tags',) _columns = ('ID', 'Branch', 'Log', 'Author', 'Date', 'Tags', 'Bookmarks') _stretchs = {'Log': 1, } _getcolumns = "getChangelogColumns" def __init__(self): """ repo is a hg repo instance """ #XXX col radius self._datacache = {} self._hasmq = False self.mqueues = [] self.wd_revs = [] self.graph = None self.rowcount = 0 self.repo = None self.show_hidden = False def setRepo(self, repo=None, branch='', fromhead=None, follow=False, closed=False): if repo is None: repo = build_repo(self.repo.ui, tounicode(self.repo.root)) self._hasmq = hasattr(self.repo, "mq") if not getattr(repo, '__hgview__', False) and self._hasmq: mqsupport.reposetup(repo.ui, repo) oldrepo = self.repo self.repo = repo if oldrepo is None or oldrepo.root != repo.root: self.load_config() self._datacache = {} try: wdctxs = self.repo[None].parents() except error.Abort: # might occur if reloading during a mq operation (or # whatever operation playing with hg history) return if self._hasmq: self.mqueues = self.repo.mq.series[:] self.wd_revs = [ctx.rev() for ctx in wdctxs] self.wd_status = [list(self.repo.status(ctx.node(), None))[:4] for ctx in wdctxs] self._user_colors = {} self._branch_colors = {} # precompute named branch color for stable value. for branch_name in chain(['default', 'stable'], sorted(allbranches(repo, True))): self.namedbranch_color(branch_name) revset_filter = _build_filter(fromhead, branch, follow, closed) grapher = revision_grapher(self.repo, revset_filter, start_rev=fromhead, show_hidden=self.show_hidden, reorder=self.reorder_changesets, show_obsolete=self.show_obsolete) self.graph = Graph(self.repo, grapher, self.max_file_size) self.rowcount = 0 self.heads = [self.repo[x].rev() for x in self.repo.heads()] self.ensureBuilt(row=self.fill_step) def ensureBuilt(self, rev=None, row=None): """ Make sure rev data is available (graph element created). """ if self.graph.isfilled(): return required = 0 buildrev = rev n = len(self.graph) if rev is not None: if n and self.graph[-1].rev <= rev: buildrev = None else: required = self.fill_step // 2 elif row is not None and row > (n - self.fill_step // 2): required = row - n + self.fill_step if required or buildrev: self.graph.build_nodes(nnodes=required, rev=buildrev) self.updateRowCount() elif row and row > self.rowcount: # asked row was already built, but views where not aware of this self.updateRowCount() elif rev is not None and rev <= self.graph[self.rowcount].rev: # asked rev was already built, but views where not aware of this self.updateRowCount() def updateRowCount(self): self.rowcount = 0 def rowCount(self, parent=None): return self.rowcount def columnCount(self, parent=None): return len(self._columns) def load_config(self): cfg = HgConfig(self.repo.ui) self._users, self._aliases = cfg.getUsers() self.dot_radius = cfg.getDotRadius(default=8) self.rowheight = cfg.getRowHeight() self.fill_step = cfg.getFillingStep() self.max_file_size = cfg.getMaxFileSize() self.hide_mq_tags = cfg.getMQHideTags() self.show_hidden = cfg.getShowHidden() self.reorder_changesets = cfg.getNonPublicOnTop() self.show_obsolete = cfg.getShowObsolete() cols = getattr(cfg, self._getcolumns)() if cols is not None: validcols = [col for col in cols if col in self._allcolumns] if len(validcols) == len(cols) and \ set(['Log', 'ID']).issubset(set(validcols)) : self._columns = tuple(validcols) @staticmethod def get_color(n, ignore=()): return [] def user_color(self, user): if user not in self._user_colors: self._user_colors[user] = self.get_color(len(self._user_colors), self._user_colors.values()) return self._user_colors[user] def user_name(self, user): return self._aliases.get(user, user) def namedbranch_color(self, branch): if branch not in self._branch_colors: self._branch_colors[branch] = self.get_color(len(self._branch_colors)) return self._branch_colors[branch] def col2x(self, col): return (1.2*self.dot_radius + 0) * col + self.dot_radius/2 + 3 def rowFromRev(self, rev): row = self.graph.index(rev) if row == -1: row = None return row def clear(self): """empty the list""" self.graph = None self._datacache = {} self.notify_data_changed() def notify_data_changed(self): pass